2025-12-04T08:53:38.3743036Z Current runner version: '2.329.0' 2025-12-04T08:53:38.3745872Z Runner name: 'linux.rocm.gpu.gfx942.1.b-gwk9b-runner-68hm6' 2025-12-04T08:53:38.3746281Z Runner group name: 'default' 2025-12-04T08:53:38.3746691Z Machine name: 'linux' 2025-12-04T08:53:38.3747883Z ##[group]GITHUB_TOKEN Permissions 2025-12-04T08:53:38.3749347Z Contents: read 2025-12-04T08:53:38.3749603Z Metadata: read 2025-12-04T08:53:38.3749824Z ##[endgroup] 2025-12-04T08:53:38.3750859Z Secret source: Actions 2025-12-04T08:53:38.3751162Z Prepare workflow directory 2025-12-04T08:53:38.3984216Z Prepare all required actions 2025-12-04T08:53:38.4003443Z Getting action download info 2025-12-04T08:53:38.9091725Z Download action repository 'pytorch/pytorch@main' (SHA:ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32) 2025-12-04T08:53:44.9771803Z Download action repository 'pytorch/test-infra@main' (SHA:39aa74d619174326f4e2fb0e216151c2f29d9ffd) 2025-12-04T08:53:46.1700313Z Download action repository 'actions/upload-artifact@ea165f8d65b6e75b540449e92b4886f43607fa02' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-12-04T08:53:47.6585336Z Download action repository 'aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722' (SHA:ececac1a45f3b08a01d2dd070d28d111c5fe6722) 2025-12-04T08:53:48.9949123Z Getting action download info 2025-12-04T08:53:49.1876082Z Download action repository 'actions/checkout@v4' (SHA:34e114876b0b11c390a56381ad16ebd13914f8d5) 2025-12-04T08:53:50.0175418Z Getting action download info 2025-12-04T08:53:50.2278946Z Download action repository 'nick-fields/retry@v3.0.0' (SHA:7152eba30c6575329ac0576536151aca5a72780e) 2025-12-04T08:53:51.1295362Z Getting action download info 2025-12-04T08:53:51.3429020Z Uses: pytorch/pytorch/.github/workflows/_rocm-test.yml@refs/heads/main (ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32) 2025-12-04T08:53:51.3431011Z ##[group] Inputs 2025-12-04T08:53:51.3431174Z build-environment: linux-jammy-rocm-py3.10 2025-12-04T08:53:51.3434391Z test-matrix: {"include": [{"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}]} 2025-12-04T08:53:51.3437797Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:53:51.3438097Z sync-tag: 2025-12-04T08:53:51.3438509Z timeout-minutes: 300 2025-12-04T08:53:51.3438630Z tests-to-include: 2025-12-04T08:53:51.3438744Z dashboard-tag: 2025-12-04T08:53:51.3438976Z disable-monitor: true 2025-12-04T08:53:51.3439107Z monitor-log-interval: 5 2025-12-04T08:53:51.3439233Z monitor-data-collect-interval: 1 2025-12-04T08:53:51.3439373Z ##[endgroup] 2025-12-04T08:53:51.3439591Z Complete job name: linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable) 2025-12-04T08:53:51.3698725Z ##[group]Run pytorch/pytorch/.github/actions/checkout-pytorch@main 2025-12-04T08:53:51.3699277Z with: 2025-12-04T08:53:51.3721584Z no-sudo: true 2025-12-04T08:53:51.3721715Z submodules: recursive 2025-12-04T08:53:51.3721828Z fetch-depth: 0 2025-12-04T08:53:51.3722152Z env: 2025-12-04T08:53:51.3722285Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:53:51.3722442Z ##[endgroup] 2025-12-04T08:53:51.3776706Z ##[group]Run echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T08:53:51.3777199Z echo "IN_CONTAINER_RUNNER=$(if [ -f /.inarc ] || [ -f /.incontainer ]; then echo true ; else echo false; fi)" >> "$GITHUB_OUTPUT" 2025-12-04T08:53:51.3784467Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:53:51.3784625Z env: 2025-12-04T08:53:51.3784720Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:53:51.3784830Z ##[endgroup] 2025-12-04T08:53:51.3914659Z ##[group]Run actions/checkout@v4 2025-12-04T08:53:51.3914791Z with: 2025-12-04T08:53:51.3914905Z ref: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:53:51.3915044Z fetch-depth: 0 2025-12-04T08:53:51.3915141Z submodules: recursive 2025-12-04T08:53:51.3915341Z show-progress: false 2025-12-04T08:53:51.3915452Z repository: pytorch/pytorch 2025-12-04T08:53:51.3915628Z token: *** 2025-12-04T08:53:51.3915721Z ssh-strict: true 2025-12-04T08:53:51.3915812Z ssh-user: git 2025-12-04T08:53:51.3915914Z persist-credentials: true 2025-12-04T08:53:51.3916028Z clean: true 2025-12-04T08:53:51.3916128Z sparse-checkout-cone-mode: true 2025-12-04T08:53:51.3916247Z fetch-tags: false 2025-12-04T08:53:51.3916341Z lfs: false 2025-12-04T08:53:51.3916437Z set-safe-directory: true 2025-12-04T08:53:51.3916541Z env: 2025-12-04T08:53:51.3916628Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:53:51.3916734Z ##[endgroup] 2025-12-04T08:53:51.4436465Z Syncing repository: pytorch/pytorch 2025-12-04T08:53:51.4437089Z ##[group]Getting Git version info 2025-12-04T08:53:51.4437255Z Working directory is '/home/runner/_work/pytorch/pytorch' 2025-12-04T08:53:51.4437507Z [command]/usr/bin/git version 2025-12-04T08:53:51.4437617Z git version 2.52.0 2025-12-04T08:53:51.4447399Z ##[endgroup] 2025-12-04T08:53:51.4455833Z Copying '/home/runner/.gitconfig' to '/home/runner/_work/_temp/558a6826-f423-4f81-9cc1-4527f51cf232/.gitconfig' 2025-12-04T08:53:51.4456677Z Temporarily overriding HOME='/home/runner/_work/_temp/558a6826-f423-4f81-9cc1-4527f51cf232' before making global git config changes 2025-12-04T08:53:51.4457494Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T08:53:51.4459743Z [command]/usr/bin/git config --global --add safe.directory /home/runner/_work/pytorch/pytorch 2025-12-04T08:53:51.4493700Z [command]/usr/bin/git config --local --get remote.origin.url 2025-12-04T08:53:51.4511341Z https://github.com/pytorch/pytorch 2025-12-04T08:53:51.4529713Z ##[group]Removing previously created refs, to avoid conflicts 2025-12-04T08:53:51.4533567Z [command]/usr/bin/git rev-parse --symbolic-full-name --verify --quiet HEAD 2025-12-04T08:53:51.4548786Z refs/heads/main 2025-12-04T08:53:51.4560765Z [command]/usr/bin/git checkout --detach 2025-12-04T08:53:52.8343757Z HEAD is now at ffd9b0fb4355 Resolve collective autotuning test failure on arm (#168919) 2025-12-04T08:53:52.8397784Z [command]/usr/bin/git branch --delete --force main 2025-12-04T08:53:52.8524549Z Deleted branch main (was ffd9b0fb4355). 2025-12-04T08:53:52.8532984Z ##[endgroup] 2025-12-04T08:53:52.8536453Z [command]/usr/bin/git submodule status 2025-12-04T08:53:52.8791251Z 7e1e1fe3858c63c251c637ae41a20de425dde96f android/libs/fbjni (v0.1.0-12-g7e1e1fe) 2025-12-04T08:53:52.8842732Z 4dfe081cf6bcd15db339cf2680b9281b8451eeb3 third_party/FP16 (4dfe081) 2025-12-04T08:53:52.8903984Z b408327ac2a15ec3e43352421954f5b1967701d1 third_party/FXdiv (b408327) 2025-12-04T08:53:52.8965954Z c07e3a0400713d546e0dea2d5466dd22ea389c73 third_party/NNPACK (c07e3a0) 2025-12-04T08:53:52.8999679Z 3ebbc93ded7285963bff932c678fa367eb393ba6 third_party/NVTX (v3.1.0-313-g3ebbc93) 2025-12-04T08:53:52.9063437Z 1d8f600fd424278486eade7ed3e877c99f0846b1 third_party/VulkanMemoryAllocator (v2.1.0-982-g1d8f600) 2025-12-04T08:53:52.9376541Z 51a0103656eff6fc9bfd39a4597923c4b542c883 third_party/XNNPACK (remotes/origin/ds/ndk-1243-g51a0103656) 2025-12-04T08:53:52.9396909Z 01aae101b9e5e94d6c16a9514c9fb8df99c93150 third_party/aiter (v0.1.1-92-g01aae101) 2025-12-04T08:53:52.9412211Z 299e5928955cc62af9968370293b916f5130916f third_party/benchmark (v1.9.3) 2025-12-04T08:53:52.9478941Z 7fe50dc3da2069d6645d9deb8c017a876472a977 third_party/composable_kernel (rocm-6.4.3-459-g7fe50dc3d) 2025-12-04T08:53:52.9571799Z 89c932f313c6437c38f2982869beacc89c2f2246 third_party/cpp-httplib (v0.26.0) 2025-12-04T08:53:52.9662282Z f858c30bcb16f8effd5ff46996f0514539e17abc third_party/cpuinfo (f858c30) 2025-12-04T08:53:52.9692574Z 0b1577c8c83401237d601d0d0db5210506705396 third_party/cudnn_frontend (v0.5-61-g0b1577c) 2025-12-04T08:53:52.9755300Z f88806b1e31dfa579842638740216dd41fc6c588 third_party/cutlass (v4.3.1) 2025-12-04T08:53:52.9775158Z c0b988d39a9e47c794d699f29930ed4d7c7e13a4 third_party/fbgemm (v1.4.0-rc1-2-gc0b988d39) 2025-12-04T08:53:52.9824673Z 979702c87a8713a8e0a5e9fee122b90d2ef13be5 third_party/flash-attention (v2.7.4) 2025-12-04T08:53:52.9841489Z a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757 third_party/flatbuffers (v24.12.23) 2025-12-04T08:53:53.0084257Z 407c905e45ad75fc29bf0f9bb7c5c2fd3475976f third_party/fmt (12.1.0) 2025-12-04T08:53:53.0147588Z 3fb5c176c17c765a3492cd2f0321b0dab712f350 third_party/gemmlowp/gemmlowp (remotes/origin/revert-87-master-135-g3fb5c17) 2025-12-04T08:53:53.0217852Z 54cbae0d3a67fa890b4c3d9ee162b7860315e341 third_party/gloo (remotes/origin/gh/c-p-i-o/1/base-37-g54cbae0) 2025-12-04T08:53:53.0355789Z 52eb8108c5bdec04579160ae17225d66034bd723 third_party/googletest (release-1.8.0-3544-g52eb8108) 2025-12-04T08:53:53.0409542Z 719d8e6cd7f7a0e01b155657526d693acf97c2b3 third_party/ideep (pytorch-rls-v3.7.1) 2025-12-04T08:53:53.0446339Z dec1d23ca65ab069d225dfe40dea14f455170959 third_party/ittapi (v3.25.5) 2025-12-04T08:53:53.0589823Z 31f85df8fbd89c188f14ef10f1ec65379786b943 third_party/kineto (heads/main) 2025-12-04T08:53:53.0612631Z d7770c89632329a9914ef1a90289917597639cbe third_party/kleidiai (v1.15.0) 2025-12-04T08:53:53.0625991Z fbd8b99c2b828428947d70fdc046bb55609be93e third_party/mimalloc (v2.2.4) 2025-12-04T08:53:53.0643364Z 55f93686c01528224f448c19128836e7df245f72 third_party/nlohmann (v3.12.0) 2025-12-04T08:53:53.0849577Z e709452ef2bbc1d113faf678c24e6d3467696e83 third_party/onnx (v1.18.0) 2025-12-04T08:53:53.0869744Z a799f4aed9c94b765dcdaabaeab7d5e7e2310878 third_party/opentelemetry-cpp (v1.14.2) 2025-12-04T08:53:53.0892024Z 0fa0ef591e38c2758e3184c6c23e497b9f732ffa third_party/pocketfft (release_for_eigen-40-g0fa0ef5) 2025-12-04T08:53:53.1097565Z d1eca4e4b421cd2997495c4b4e65cea6be4e9b8a third_party/protobuf (v3.7.0-rc.2-1279-gd1eca4e4b) 2025-12-04T08:53:53.1141493Z 072586a71b55b7f8c584153d223e95687148a900 third_party/psimd (heads/master) 2025-12-04T08:53:53.1185366Z 4fe0e1e183925bf8cfa6aae24237e724a96479b8 third_party/pthreadpool (0.1-144-g4fe0e1e) 2025-12-04T08:53:53.1204183Z f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8 third_party/pybind11 (v3.0.1) 2025-12-04T08:53:53.1268075Z f45429b087dd7d5bc78bb40dc7cf06425c252d67 third_party/python-peachpy (remotes/origin/pre-generated) 2025-12-04T08:53:53.1321546Z 5a1d179df9cf652951b59010a2d2075372d67f68 third_party/sleef (3.8) 2025-12-04T08:53:53.1373283Z 2b4cd91092d335a697416b2a3cb398283246849d third_party/tensorpipe (heads/main) 2025-12-04T08:53:53.1382554Z ##[group]Cleaning the repository 2025-12-04T08:53:53.1387208Z [command]/usr/bin/git clean -ffdx 2025-12-04T08:53:53.1502348Z [command]/usr/bin/git reset --hard HEAD 2025-12-04T08:53:53.2102847Z HEAD is now at ffd9b0fb4355 Resolve collective autotuning test failure on arm (#168919) 2025-12-04T08:53:53.2146326Z ##[endgroup] 2025-12-04T08:53:53.2148858Z ##[group]Disabling automatic garbage collection 2025-12-04T08:53:53.2151632Z [command]/usr/bin/git config --local gc.auto 0 2025-12-04T08:53:53.2170911Z ##[endgroup] 2025-12-04T08:53:53.2171108Z ##[group]Setting up auth 2025-12-04T08:53:53.2174221Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T08:53:53.2191392Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T08:53:53.2360809Z Entering 'android/libs/fbjni' 2025-12-04T08:53:53.2384958Z Entering 'third_party/FP16' 2025-12-04T08:53:53.2414704Z Entering 'third_party/FXdiv' 2025-12-04T08:53:53.2437599Z Entering 'third_party/NNPACK' 2025-12-04T08:53:53.2460544Z Entering 'third_party/NVTX' 2025-12-04T08:53:53.2481915Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:53:53.2509849Z Entering 'third_party/XNNPACK' 2025-12-04T08:53:53.2538467Z Entering 'third_party/aiter' 2025-12-04T08:53:53.2565330Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:53:53.2593954Z Entering 'third_party/benchmark' 2025-12-04T08:53:53.2614843Z Entering 'third_party/composable_kernel' 2025-12-04T08:53:53.2641673Z Entering 'third_party/cpp-httplib' 2025-12-04T08:53:53.2663170Z Entering 'third_party/cpuinfo' 2025-12-04T08:53:53.2682591Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:53:53.2701977Z Entering 'third_party/cutlass' 2025-12-04T08:53:53.2726468Z Entering 'third_party/fbgemm' 2025-12-04T08:53:53.2755843Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:53:53.2782347Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:53:53.2810395Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:53:53.2834867Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:53:53.2862526Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:53:53.2889983Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:53:53.2910640Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:53:53.2933534Z Entering 'third_party/flash-attention' 2025-12-04T08:53:53.2954474Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:53:53.2987498Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:53:53.3013318Z Entering 'third_party/flatbuffers' 2025-12-04T08:53:53.3039705Z Entering 'third_party/fmt' 2025-12-04T08:53:53.3061148Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:53:53.3088376Z Entering 'third_party/gloo' 2025-12-04T08:53:53.3109377Z Entering 'third_party/googletest' 2025-12-04T08:53:53.3136933Z Entering 'third_party/ideep' 2025-12-04T08:53:53.3160028Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:53:53.3185960Z Entering 'third_party/ittapi' 2025-12-04T08:53:53.3208598Z Entering 'third_party/kineto' 2025-12-04T08:53:53.3230421Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:53:53.3252569Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:53:53.3274590Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:53:53.3296162Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:53:53.3317712Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:53:53.3337999Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:53:53.3361756Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:53:53.3381696Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:53:53.3407248Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:53:53.3429340Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:53:53.3448825Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:53:53.3469042Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:53:53.3497195Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:53:53.3522933Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:53:53.3552395Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:53:53.3582268Z Entering 'third_party/kleidiai' 2025-12-04T08:53:53.3629023Z Entering 'third_party/mimalloc' 2025-12-04T08:53:53.3664312Z Entering 'third_party/nlohmann' 2025-12-04T08:53:53.3691752Z Entering 'third_party/onnx' 2025-12-04T08:53:53.3723588Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:53:53.3758294Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:53:53.3785801Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:53:53.3809088Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:53:53.3835142Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:53:53.3858055Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:53:53.3880924Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:53:53.3907607Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:53:53.3928978Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:53:53.3953161Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:53:53.3975842Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:53:53.4008370Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:53:53.4038815Z Entering 'third_party/pocketfft' 2025-12-04T08:53:53.4061157Z Entering 'third_party/protobuf' 2025-12-04T08:53:53.4089924Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:53:53.4111898Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:53:53.4136916Z Entering 'third_party/psimd' 2025-12-04T08:53:53.4160094Z Entering 'third_party/pthreadpool' 2025-12-04T08:53:53.4183490Z Entering 'third_party/pybind11' 2025-12-04T08:53:53.4208449Z Entering 'third_party/python-peachpy' 2025-12-04T08:53:53.4242336Z Entering 'third_party/sleef' 2025-12-04T08:53:53.4273521Z Entering 'third_party/tensorpipe' 2025-12-04T08:53:53.4298250Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:53:53.4320758Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:53:53.4344535Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:53:53.4365140Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:53:53.4387842Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:53:53.4435405Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T08:53:53.4455318Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T08:53:53.4604948Z Entering 'android/libs/fbjni' 2025-12-04T08:53:53.4627527Z Entering 'third_party/FP16' 2025-12-04T08:53:53.4649441Z Entering 'third_party/FXdiv' 2025-12-04T08:53:53.4671508Z Entering 'third_party/NNPACK' 2025-12-04T08:53:53.4693382Z Entering 'third_party/NVTX' 2025-12-04T08:53:53.4714942Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:53:53.4737864Z Entering 'third_party/XNNPACK' 2025-12-04T08:53:53.4765245Z Entering 'third_party/aiter' 2025-12-04T08:53:53.4789379Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:53:53.4813735Z Entering 'third_party/benchmark' 2025-12-04T08:53:53.4834921Z Entering 'third_party/composable_kernel' 2025-12-04T08:53:53.4858664Z Entering 'third_party/cpp-httplib' 2025-12-04T08:53:53.4881726Z Entering 'third_party/cpuinfo' 2025-12-04T08:53:53.4904419Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:53:53.4929797Z Entering 'third_party/cutlass' 2025-12-04T08:53:53.4958434Z Entering 'third_party/fbgemm' 2025-12-04T08:53:53.4981520Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:53:53.5004384Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:53:53.5028407Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:53:53.5049774Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:53:53.5075339Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:53:53.5094362Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:53:53.5119992Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:53:53.5143832Z Entering 'third_party/flash-attention' 2025-12-04T08:53:53.5170357Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:53:53.5192204Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:53:53.5225042Z Entering 'third_party/flatbuffers' 2025-12-04T08:53:53.5248787Z Entering 'third_party/fmt' 2025-12-04T08:53:53.5273885Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:53:53.5295438Z Entering 'third_party/gloo' 2025-12-04T08:53:53.5316676Z Entering 'third_party/googletest' 2025-12-04T08:53:53.5339561Z Entering 'third_party/ideep' 2025-12-04T08:53:53.5361128Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:53:53.5396477Z Entering 'third_party/ittapi' 2025-12-04T08:53:53.5418121Z Entering 'third_party/kineto' 2025-12-04T08:53:53.5439523Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:53:53.5472752Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:53:53.5498213Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:53:53.5522977Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:53:53.5543230Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:53:53.5568957Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:53:53.5591465Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:53:53.5613554Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:53:53.5634000Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:53:53.5654088Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:53:53.5674619Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:53:53.5695092Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:53:53.5716277Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:53:53.5739931Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:53:53.5763935Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:53:53.5785541Z Entering 'third_party/kleidiai' 2025-12-04T08:53:53.5808217Z Entering 'third_party/mimalloc' 2025-12-04T08:53:53.5832004Z Entering 'third_party/nlohmann' 2025-12-04T08:53:53.5853898Z Entering 'third_party/onnx' 2025-12-04T08:53:53.5886541Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:53:53.5911556Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:53:53.5932251Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:53:53.5952491Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:53:53.5973292Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:53:53.5995812Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:53:53.6024269Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:53:53.6045301Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:53:53.6065732Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:53:53.6084248Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:53:53.6112858Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:53:53.6134863Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:53:53.6163523Z Entering 'third_party/pocketfft' 2025-12-04T08:53:53.6184539Z Entering 'third_party/protobuf' 2025-12-04T08:53:53.6205809Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:53:53.6230210Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:53:53.6252385Z Entering 'third_party/psimd' 2025-12-04T08:53:53.6278203Z Entering 'third_party/pthreadpool' 2025-12-04T08:53:53.6300487Z Entering 'third_party/pybind11' 2025-12-04T08:53:53.6322284Z Entering 'third_party/python-peachpy' 2025-12-04T08:53:53.6341469Z Entering 'third_party/sleef' 2025-12-04T08:53:53.6361944Z Entering 'third_party/tensorpipe' 2025-12-04T08:53:53.6382361Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:53:53.6405907Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:53:53.6430973Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:53:53.6455413Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:53:53.6476810Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:53:53.6515005Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.6537158Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T08:53:53.6688981Z Entering 'android/libs/fbjni' 2025-12-04T08:53:53.6698369Z file:/home/runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T08:53:53.6707027Z Entering 'third_party/FP16' 2025-12-04T08:53:53.6716518Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T08:53:53.6725160Z Entering 'third_party/FXdiv' 2025-12-04T08:53:53.6734842Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T08:53:53.6743514Z Entering 'third_party/NNPACK' 2025-12-04T08:53:53.6754192Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T08:53:53.6762846Z Entering 'third_party/NVTX' 2025-12-04T08:53:53.6772146Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T08:53:53.6781035Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:53:53.6792622Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T08:53:53.6805433Z Entering 'third_party/XNNPACK' 2025-12-04T08:53:53.6817386Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T08:53:53.6831669Z Entering 'third_party/aiter' 2025-12-04T08:53:53.6841576Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T08:53:53.6850375Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:53:53.6860430Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T08:53:53.6872261Z Entering 'third_party/benchmark' 2025-12-04T08:53:53.6882417Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:53:53.6892087Z Entering 'third_party/composable_kernel' 2025-12-04T08:53:53.6902697Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T08:53:53.6916246Z Entering 'third_party/cpp-httplib' 2025-12-04T08:53:53.6930911Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T08:53:53.6940750Z Entering 'third_party/cpuinfo' 2025-12-04T08:53:53.6952913Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T08:53:53.6961896Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:53:53.6972094Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T08:53:53.6981479Z Entering 'third_party/cutlass' 2025-12-04T08:53:53.6991213Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T08:53:53.7004431Z Entering 'third_party/fbgemm' 2025-12-04T08:53:53.7015213Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T08:53:53.7025238Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:53:53.7035402Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T08:53:53.7042359Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:53:53.7051310Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T08:53:53.7062260Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:53:53.7082255Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T08:53:53.7091097Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:53:53.7103443Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T08:53:53.7122481Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:53:53.7132931Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T08:53:53.7141784Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:53:53.7151269Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T08:53:53.7163225Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:53:53.7172749Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T08:53:53.7184993Z Entering 'third_party/flash-attention' 2025-12-04T08:53:53.7195359Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T08:53:53.7204244Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:53:53.7213314Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T08:53:53.7223349Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:53:53.7232924Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T08:53:53.7247864Z Entering 'third_party/flatbuffers' 2025-12-04T08:53:53.7259581Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T08:53:53.7270247Z Entering 'third_party/fmt' 2025-12-04T08:53:53.7283201Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T08:53:53.7292985Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:53:53.7305595Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T08:53:53.7314482Z Entering 'third_party/gloo' 2025-12-04T08:53:53.7324591Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T08:53:53.7333452Z Entering 'third_party/googletest' 2025-12-04T08:53:53.7345306Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:53:53.7354217Z Entering 'third_party/ideep' 2025-12-04T08:53:53.7364537Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T08:53:53.7373561Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:53:53.7387127Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T08:53:53.7399838Z Entering 'third_party/ittapi' 2025-12-04T08:53:53.7412966Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T08:53:53.7424353Z Entering 'third_party/kineto' 2025-12-04T08:53:53.7435279Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T08:53:53.7446074Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:53:53.7462417Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T08:53:53.7480175Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:53:53.7493154Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T08:53:53.7509094Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:53:53.7523119Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T08:53:53.7544618Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:53:53.7561704Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T08:53:53.7576121Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:53:53.7588803Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T08:53:53.7597234Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:53:53.7616473Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T08:53:53.7631180Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:53:53.7641133Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T08:53:53.7655974Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:53:53.7668190Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:53:53.7680722Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:53:53.7691897Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T08:53:53.7703950Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:53:53.7714789Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T08:53:53.7724605Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:53:53.7736133Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T08:53:53.7750907Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:53:53.7762156Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T08:53:53.7772753Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:53:53.7782870Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T08:53:53.7798134Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:53:53.7808867Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T08:53:53.7818808Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:53:53.7828342Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T08:53:53.7837971Z Entering 'third_party/kleidiai' 2025-12-04T08:53:53.7849631Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T08:53:53.7859866Z Entering 'third_party/mimalloc' 2025-12-04T08:53:53.7871940Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T08:53:53.7881080Z Entering 'third_party/nlohmann' 2025-12-04T08:53:53.7891365Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T08:53:53.7902190Z Entering 'third_party/onnx' 2025-12-04T08:53:53.7912179Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T08:53:53.7926975Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:53:53.7936533Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:53:53.7948323Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:53:53.7958685Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T08:53:53.7968153Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:53:53.7980890Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:53:53.7990285Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:53:53.8005774Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:53:53.8016299Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:53:53.8030827Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T08:53:53.8047552Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:53:53.8059220Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T08:53:53.8072106Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:53:53.8092249Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T08:53:53.8100720Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:53:53.8114677Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T08:53:53.8124081Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:53:53.8136975Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T08:53:53.8149409Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:53:53.8168454Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T08:53:53.8178825Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:53:53.8194277Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T08:53:53.8205607Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:53:53.8216208Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T08:53:53.8233724Z Entering 'third_party/pocketfft' 2025-12-04T08:53:53.8244960Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T08:53:53.8255325Z Entering 'third_party/protobuf' 2025-12-04T08:53:53.8265427Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T08:53:53.8276276Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:53:53.8287938Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:53:53.8296889Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:53:53.8305881Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:53:53.8315625Z Entering 'third_party/psimd' 2025-12-04T08:53:53.8325411Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T08:53:53.8334058Z Entering 'third_party/pthreadpool' 2025-12-04T08:53:53.8343954Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T08:53:53.8352411Z Entering 'third_party/pybind11' 2025-12-04T08:53:53.8361956Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:53:53.8371620Z Entering 'third_party/python-peachpy' 2025-12-04T08:53:53.8381810Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T08:53:53.8392266Z Entering 'third_party/sleef' 2025-12-04T08:53:53.8402798Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T08:53:53.8412953Z Entering 'third_party/tensorpipe' 2025-12-04T08:53:53.8423892Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T08:53:53.8433248Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:53:53.8442369Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:53:53.8450838Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:53:53.8461740Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T08:53:53.8478327Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:53:53.8487865Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T08:53:53.8496185Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:53:53.8504845Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:53:53.8512642Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:53:53.8527659Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T08:53:53.8551094Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8570347Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8585638Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8600411Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8615095Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8630376Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8644488Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8658413Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8672325Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8685877Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8702044Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8715573Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8729234Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8742510Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8756471Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8769810Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8786924Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8797499Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8811606Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8824720Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8839990Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8853158Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8867056Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8880399Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8894751Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8908639Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8921112Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8933548Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8954038Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8967489Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8981794Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.8994833Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9009479Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9024225Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9042401Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9056197Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9071763Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9085521Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9102353Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9115837Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9130399Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9144087Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9156652Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9169230Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9184216Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9197898Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9212118Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9225688Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9238045Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9250828Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9263093Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9276061Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9290807Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9304620Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9319725Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9338683Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9354448Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9369343Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9383942Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9399548Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9413446Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9427986Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9442189Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9456190Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9471983Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9485739Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9500601Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9519892Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9534278Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9547298Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9560034Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9572553Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9587447Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9602004Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9616466Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9629323Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9643884Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9659279Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9672484Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9696100Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9714122Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:53:53.9733513Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T08:53:53.9755770Z ##[endgroup] 2025-12-04T08:53:53.9755955Z ##[group]Fetching the repository 2025-12-04T08:53:53.9759602Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-12-04T08:53:54.6703445Z From https://github.com/pytorch/pytorch 2025-12-04T08:53:54.6703858Z - [deleted] (none) -> ciflow/inductor/160174 2025-12-04T08:53:54.6704231Z - [deleted] (none) -> ciflow/trunk/160174 2025-12-04T08:53:58.7241457Z * [new branch] 2.6.0.dev20241004+ -> origin/2.6.0.dev20241004+ 2025-12-04T08:53:58.7241853Z * [new branch] 2.9.1 -> origin/2.9.1 2025-12-04T08:53:58.7242293Z * [new branch] AaronWang04_addmmfusion_perftest -> origin/AaronWang04_addmmfusion_perftest 2025-12-04T08:53:58.7242708Z * [new branch] Flamefire-patch-1 -> origin/Flamefire-patch-1 2025-12-04T08:53:58.7243086Z * [new branch] HDCharles-2.6.0-release-notes -> origin/HDCharles-2.6.0-release-notes 2025-12-04T08:53:58.7243482Z * [new branch] HOPrintFunc -> origin/HOPrintFunc 2025-12-04T08:53:58.7243882Z * [new branch] IvanKobzarev/stack/1 -> origin/IvanKobzarev/stack/1 2025-12-04T08:53:58.7244215Z * [new branch] NicoshevSVE128 -> origin/NicoshevSVE128 2025-12-04T08:53:58.7244553Z * [new branch] PR-AOTInductorNoneBug -> origin/PR-AOTInductorNoneBug 2025-12-04T08:53:58.7244926Z * [new branch] PR-AOTInductorNoneBugFix -> origin/PR-AOTInductorNoneBugFix 2025-12-04T08:53:58.7245291Z * [new branch] PR-FixConfigsIssue -> origin/PR-FixConfigsIssue 2025-12-04T08:53:58.7246248Z * [new branch] PR-NoneBugFix-viable -> origin/PR-NoneBugFix-viable 2025-12-04T08:53:58.7246576Z * [new branch] PR-ResetToZero -> origin/PR-ResetToZero 2025-12-04T08:53:58.7247040Z * [new branch] Update-Flash-Packaging -> origin/Update-Flash-Packaging 2025-12-04T08:53:58.7247388Z * [new branch] VLA_exp -> origin/VLA_exp 2025-12-04T08:53:58.7247695Z * [new branch] activation_bench -> origin/activation_bench 2025-12-04T08:53:58.7248015Z * [new branch] addmm-heuristic -> origin/addmm-heuristic 2025-12-04T08:53:58.7248355Z * [new branch] adi/onednn_aarch64 -> origin/adi/onednn_aarch64 2025-12-04T08:53:58.7248663Z * [new branch] adi/test -> origin/adi/test 2025-12-04T08:53:58.7248955Z * [new branch] adi/test_bgemm -> origin/adi/test_bgemm 2025-12-04T08:53:58.7249267Z * [new branch] adi/test_m8g -> origin/adi/test_m8g 2025-12-04T08:53:58.7249571Z * [new branch] adi/test_onednn -> origin/adi/test_onednn 2025-12-04T08:53:58.7249892Z * [new branch] adi/test_onednn_v3.9 -> origin/adi/test_onednn_v3.9 2025-12-04T08:53:58.7250233Z * [new branch] adi/test_presve_change -> origin/adi/test_presve_change 2025-12-04T08:53:58.7250684Z * [new branch] adi/test_timm -> origin/adi/test_timm 2025-12-04T08:53:58.7251015Z * [new branch] adi/testpresve_change -> origin/adi/testpresve_change 2025-12-04T08:53:58.7251366Z * [new branch] aditew01/test/vec_bf16 -> origin/aditew01/test/vec_bf16 2025-12-04T08:53:58.7251719Z * [new branch] ah-globalfeedback-hook -> origin/ah-globalfeedback-hook 2025-12-04T08:53:58.7252098Z * [new branch] albanD-patch-1 -> origin/albanD-patch-1 2025-12-04T08:53:58.7252438Z * [new branch] also-surround-shimh -> origin/also-surround-shimh 2025-12-04T08:53:58.7252783Z * [new branch] angelayi/aot_compile -> origin/angelayi/aot_compile 2025-12-04T08:53:58.7253166Z * [new branch] angelayi/aoti_additional_files -> origin/angelayi/aoti_additional_files 2025-12-04T08:53:58.7253547Z * [new branch] angelayi/benchmark -> origin/angelayi/benchmark 2025-12-04T08:53:58.7253848Z * [new branch] angelayi/change_pytree_serialization -> origin/angelayi/change_pytree_serialization 2025-12-04T08:53:58.7254158Z * [new branch] angelayi/cpp_loader -> origin/angelayi/cpp_loader 2025-12-04T08:53:58.7254415Z * [new branch] angelayi/inductor_const -> origin/angelayi/inductor_const 2025-12-04T08:53:58.7254660Z * [new branch] angelayi/lstm -> origin/angelayi/lstm 2025-12-04T08:53:58.7254910Z * [new branch] angelayi/no_so_weight -> origin/angelayi/no_so_weight 2025-12-04T08:53:58.7255163Z * [new branch] angelayi/scan_layers -> origin/angelayi/scan_layers 2025-12-04T08:53:58.7255397Z * [new branch] angelayi/side_eff -> origin/angelayi/side_eff 2025-12-04T08:53:58.7255638Z * [new branch] angelayi/state_dict -> origin/angelayi/state_dict 2025-12-04T08:53:58.7255894Z * [new branch] angelayi/symint_input -> origin/angelayi/symint_input 2025-12-04T08:53:58.7256139Z * [new branch] angelayi/symm_mem -> origin/angelayi/symm_mem 2025-12-04T08:53:58.7256402Z * [new branch] angelayi/test_cpp -> origin/angelayi/test_cpp 2025-12-04T08:53:58.7256643Z * [new branch] angelayi/torch_size -> origin/angelayi/torch_size 2025-12-04T08:53:58.7256924Z * [new branch] annotate_assert -> origin/annotate_assert 2025-12-04T08:53:58.7257176Z * [new branch] annotate_fallback_kernel -> origin/annotate_fallback_kernel 2025-12-04T08:53:58.7257565Z * [new branch] annotation_deepcopy -> origin/annotation_deepcopy 2025-12-04T08:53:58.7257807Z * [new branch] annotation_dynamo -> origin/annotation_dynamo 2025-12-04T08:53:58.7258046Z * [new branch] aot_eager_stack_trace -> origin/aot_eager_stack_trace 2025-12-04T08:53:58.7258293Z * [new branch] aoti-cuda-alloc -> origin/aoti-cuda-alloc 2025-12-04T08:53:58.7258529Z * [new branch] aoti_const_device -> origin/aoti_const_device 2025-12-04T08:53:58.7258775Z * [new branch] aoti_fqn_name_interface -> origin/aoti_fqn_name_interface 2025-12-04T08:53:58.7259057Z * [new branch] aoti_package_weights_binary -> origin/aoti_package_weights_binary 2025-12-04T08:53:58.7259319Z * [new branch] aoti_target_windows -> origin/aoti_target_windows 2025-12-04T08:53:58.7259613Z * [new branch] arsh/feat/inductor_check_profiling -> origin/arsh/feat/inductor_check_profiling 2025-12-04T08:53:58.7259887Z * [new branch] async_tp -> origin/async_tp 2025-12-04T08:53:58.7260151Z * [new branch] atalman-inductor-perf-cu124 -> origin/atalman-inductor-perf-cu124 2025-12-04T08:53:58.7260473Z * [new branch] atalman-inductor-perf-cu124.1 -> origin/atalman-inductor-perf-cu124.1 2025-12-04T08:53:58.7260807Z * [new branch] atalman-patch-2 -> origin/atalman-patch-2 2025-12-04T08:53:58.7261049Z * [new branch] atalman-patch-3 -> origin/atalman-patch-3 2025-12-04T08:53:58.7261276Z * [new branch] atalman-patch-4 -> origin/atalman-patch-4 2025-12-04T08:53:58.7261512Z * [new branch] atalman-patch-5 -> origin/atalman-patch-5 2025-12-04T08:53:58.7261739Z * [new branch] atalman-patch-6 -> origin/atalman-patch-6 2025-12-04T08:53:58.7261976Z * [new branch] atalman-patch-7 -> origin/atalman-patch-7 2025-12-04T08:53:58.7262207Z * [new branch] atalman-patch-8 -> origin/atalman-patch-8 2025-12-04T08:53:58.7262457Z * [new branch] atalman_inductor_2.3.1 -> origin/atalman_inductor_2.3.1 2025-12-04T08:53:58.7262712Z * [new branch] atalman_inductor_2.4.0 -> origin/atalman_inductor_2.4.0 2025-12-04T08:53:58.7262973Z * [new branch] atalman_inductor_2.4.x -> origin/atalman_inductor_2.4.x 2025-12-04T08:53:58.7263263Z * [new branch] attention_benchmarking_clean -> origin/attention_benchmarking_clean 2025-12-04T08:53:58.7263596Z * [new branch] bahuang/dt_fix_scalar_add -> origin/bahuang/dt_fix_scalar_add 2025-12-04T08:53:58.7263812Z * [new branch] bahuang/fix_debug_mode -> origin/bahuang/fix_debug_mode 2025-12-04T08:53:58.7264012Z * [new branch] bahuang/fix_expand -> origin/bahuang/fix_expand 2025-12-04T08:53:58.7264204Z * [new branch] bahuang/test -> origin/bahuang/test 2025-12-04T08:53:58.7264385Z * [new branch] base/1.5 -> origin/base/1.5 2025-12-04T08:53:58.7264695Z * [new branch] batching_sdpa_efficient_attention -> origin/batching_sdpa_efficient_attention 2025-12-04T08:53:58.7264927Z * [new branch] bench_scaled_mm_ops -> origin/bench_scaled_mm_ops 2025-12-04T08:53:58.7265130Z * [new branch] benchmark-updates -> origin/benchmark-updates 2025-12-04T08:53:58.7265340Z * [new branch] benchmarking-script -> origin/benchmarking-script 2025-12-04T08:53:58.7265544Z * [new branch] bertmaher/pinbump26 -> origin/bertmaher/pinbump26 2025-12-04T08:53:58.7265744Z * [new branch] bertrand/cutlass -> origin/bertrand/cutlass 2025-12-04T08:53:58.7265945Z * [new branch] bf/bug-static-input -> origin/bf/bug-static-input 2025-12-04T08:53:58.7266207Z * [new branch] bf/cg-backend -> origin/bf/cg-backend 2025-12-04T08:53:58.7266396Z * [new branch] bf/cg-nccl-test -> origin/bf/cg-nccl-test 2025-12-04T08:53:58.7266593Z * [new branch] bf/cg-remove-check -> origin/bf/cg-remove-check 2025-12-04T08:53:58.7266844Z * [new branch] bf/clean-torchbench-hf -> origin/bf/clean-torchbench-hf 2025-12-04T08:53:58.7267055Z * [new branch] bf/combo-debug-log -> origin/bf/combo-debug-log 2025-12-04T08:53:58.7267252Z * [new branch] bf/cudagraph -> origin/bf/cudagraph 2025-12-04T08:53:58.7267494Z * [new branch] bf/cudagraph-disable-input-mutation -> origin/bf/cudagraph-disable-input-mutation 2025-12-04T08:53:58.7267881Z * [new branch] bf/cudagraph-enable-input-mutation-support-benchmark -> origin/bf/cudagraph-enable-input-mutation-support-benchmark 2025-12-04T08:53:58.7268216Z * [new branch] bf/cudagraph-partition -> origin/bf/cudagraph-partition 2025-12-04T08:53:58.7268433Z * [new branch] bf/donated-buffer-bench -> origin/bf/donated-buffer-bench 2025-12-04T08:53:58.7268651Z * [new branch] bf/dynamo-partition -> origin/bf/dynamo-partition 2025-12-04T08:53:58.7268841Z * [new branch] bf/lite -> origin/bf/lite 2025-12-04T08:53:58.7269084Z * [new branch] bf/pa-non-divisible -> origin/bf/pa-non-divisible 2025-12-04T08:53:58.7269329Z * [new branch] bf/partition-cache-free-symbols -> origin/bf/partition-cache-free-symbols 2025-12-04T08:53:58.7269586Z * [new branch] bf/partition-memory-plan -> origin/bf/partition-memory-plan 2025-12-04T08:53:58.7269815Z * [new branch] bf/partition-move-cpu -> origin/bf/partition-move-cpu 2025-12-04T08:53:58.7270047Z * [new branch] bf/partition-view-fallback -> origin/bf/partition-view-fallback 2025-12-04T08:53:58.7270286Z * [new branch] bf/remove-check-55b0c39d -> origin/bf/remove-check-55b0c39d 2025-12-04T08:53:58.7270501Z * [new branch] bf/timm-nov-26-2025 -> origin/bf/timm-nov-26-2025 2025-12-04T08:53:58.7270731Z * [new branch] bf/transformer-pin-4-57-3 -> origin/bf/transformer-pin-4-57-3 2025-12-04T08:53:58.7270971Z * [new branch] bisect_perf_hf_T5_3acc6eac492 -> origin/bisect_perf_hf_T5_3acc6eac492 2025-12-04T08:53:58.7271211Z * [new branch] bisect_perf_hf_T5_3fcf66f61fb -> origin/bisect_perf_hf_T5_3fcf66f61fb 2025-12-04T08:53:58.7271447Z * [new branch] bisect_perf_hf_T5_4009d154129 -> origin/bisect_perf_hf_T5_4009d154129 2025-12-04T08:53:58.7271673Z * [new branch] bisect_perf_hf_T5_40d0740e73d -> origin/bisect_perf_hf_T5_40d0740e73d 2025-12-04T08:53:58.7271896Z * [new branch] bisect_perf_hf_T5_5268754e -> origin/bisect_perf_hf_T5_5268754e 2025-12-04T08:53:58.7272131Z * [new branch] bisect_perf_hf_T5_7d89a8d385c -> origin/bisect_perf_hf_T5_7d89a8d385c 2025-12-04T08:53:58.7272357Z * [new branch] bisect_perf_hf_T5_b7a25c1ee7c -> origin/bisect_perf_hf_T5_b7a25c1ee7c 2025-12-04T08:53:58.7272588Z * [new branch] bisect_perf_hf_T5_c25b201583f -> origin/bisect_perf_hf_T5_c25b201583f 2025-12-04T08:53:58.7272825Z * [new branch] bisect_perf_hf_T5_c93e57efac0 -> origin/bisect_perf_hf_T5_c93e57efac0 2025-12-04T08:53:58.7273052Z * [new branch] bisect_perf_hf_T5_ca9813ea149 -> origin/bisect_perf_hf_T5_ca9813ea149 2025-12-04T08:53:58.7273279Z * [new branch] bisect_perf_hf_T5_d65f194a -> origin/bisect_perf_hf_T5_d65f194a 2025-12-04T08:53:58.7273501Z * [new branch] bisect_perf_hf_T5_da94ab0b -> origin/bisect_perf_hf_T5_da94ab0b 2025-12-04T08:53:58.7273724Z * [new branch] bisect_perf_hf_T5_da94ab0b_new -> origin/bisect_perf_hf_T5_da94ab0b_new 2025-12-04T08:53:58.7273993Z * [new branch] bisect_perf_hf_T5_db4e8a1d8a8 -> origin/bisect_perf_hf_T5_db4e8a1d8a8 2025-12-04T08:53:58.7274223Z * [new branch] bisect_perf_hf_T5_e0d97e936a2 -> origin/bisect_perf_hf_T5_e0d97e936a2 2025-12-04T08:53:58.7274444Z * [new branch] bisect_perf_hf_T5_f23621ec563 -> origin/bisect_perf_hf_T5_f23621ec563 2025-12-04T08:53:58.7274658Z * [new branch] brister/fx_device_type -> origin/brister/fx_device_type 2025-12-04T08:53:58.7274883Z * [new branch] brister/test_inductor_all_fx -> origin/brister/test_inductor_all_fx 2025-12-04T08:53:58.7275144Z * [new branch] brister/tiled_reduction_no_numel_check -> origin/brister/tiled_reduction_no_numel_check 2025-12-04T08:53:58.7275379Z * [new branch] bwd-backup -> origin/bwd-backup 2025-12-04T08:53:58.7275554Z * [new branch] c57382a49 -> origin/c57382a49 2025-12-04T08:53:58.7275719Z * [new branch] ca_0431d47eaa -> origin/ca_0431d47eaa 2025-12-04T08:53:58.7275899Z * [new branch] ca_fix_0431d47eaa -> origin/ca_fix_0431d47eaa 2025-12-04T08:53:58.7276114Z * [new branch] camyllh/test_setup_hooks_push -> origin/camyllh/test_setup_hooks_push 2025-12-04T08:53:58.7276322Z * [new branch] cccclai-patch-1 -> origin/cccclai-patch-1 2025-12-04T08:53:58.7276596Z * [new branch] cherry-pick-159969-by-pytorch_bot_bot_ -> origin/cherry-pick-159969-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7276946Z * [new branch] cherry-pick-160586-by-pytorch_bot_bot_ -> origin/cherry-pick-160586-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7277224Z * [new branch] cherry-pick-162208-by-pytorch_bot_bot_ -> origin/cherry-pick-162208-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7277503Z * [new branch] cherry-pick-163169-by-pytorch_bot_bot_ -> origin/cherry-pick-163169-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7277784Z * [new branch] cherry-pick-165086-by-pytorch_bot_bot_ -> origin/cherry-pick-165086-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7278065Z * [new branch] cherry-pick-165514-by-pytorch_bot_bot_ -> origin/cherry-pick-165514-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7278343Z * [new branch] cherry-pick-165601-by-pytorch_bot_bot_ -> origin/cherry-pick-165601-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7278626Z * [new branch] cherry-pick-165667-by-pytorch_bot_bot_ -> origin/cherry-pick-165667-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7278917Z * [new branch] cherry-pick-165815-by-pytorch_bot_bot_ -> origin/cherry-pick-165815-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7279196Z * [new branch] cherry-pick-165922-by-pytorch_bot_bot_ -> origin/cherry-pick-165922-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7279477Z * [new branch] cherry-pick-166148-by-pytorch_bot_bot_ -> origin/cherry-pick-166148-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7279758Z * [new branch] cherry-pick-166181-by-pytorch_bot_bot_ -> origin/cherry-pick-166181-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7280036Z * [new branch] cherry-pick-166404-by-pytorch_bot_bot_ -> origin/cherry-pick-166404-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7280314Z * [new branch] cherry-pick-166427-by-pytorch_bot_bot_ -> origin/cherry-pick-166427-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7280596Z * [new branch] cherry-pick-166480-by-pytorch_bot_bot_ -> origin/cherry-pick-166480-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7280874Z * [new branch] cherry-pick-166570-by-pytorch_bot_bot_ -> origin/cherry-pick-166570-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7281153Z * [new branch] cherry-pick-166993-by-pytorch_bot_bot_ -> origin/cherry-pick-166993-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7281432Z * [new branch] cherry-pick-167111-by-pytorch_bot_bot_ -> origin/cherry-pick-167111-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7281750Z * [new branch] cherry-pick-167478-by-pytorch_bot_bot_ -> origin/cherry-pick-167478-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7281992Z * [new branch] cherry_pick_166036_166040 -> origin/cherry_pick_166036_166040 2025-12-04T08:53:58.7282189Z * [new branch] cherry_pick_166457 -> origin/cherry_pick_166457 2025-12-04T08:53:58.7282377Z * [new branch] cherrypick_166338 -> origin/cherrypick_166338 2025-12-04T08:53:58.7282564Z * [new branch] cherrypick_166458 -> origin/cherrypick_166458 2025-12-04T08:53:58.7282745Z * [new branch] cherrypick_166586 -> origin/cherrypick_166586 2025-12-04T08:53:58.7282929Z * [new branch] cherrypick_166956 -> origin/cherrypick_166956 2025-12-04T08:53:58.7283109Z * [new branch] ci_attn -> origin/ci_attn 2025-12-04T08:53:58.7283281Z * [new branch] codex-testing -> origin/codex-testing 2025-12-04T08:53:58.7283553Z * [new branch] codex/add-check_memory_overlap-helper-functions -> origin/codex/add-check_memory_overlap-helper-functions 2025-12-04T08:53:58.7283865Z * [new branch] codex/fix-issue-121219-in-pytorch -> origin/codex/fix-issue-121219-in-pytorch 2025-12-04T08:53:58.7284211Z * [new branch] codex/investigate-segfaults-in-get_tensor_storage_id -> origin/codex/investigate-segfaults-in-get_tensor_storage_id 2025-12-04T08:53:58.7284584Z * [new branch] codex/refactor-lintrunner-config-to-use-uv-run -> origin/codex/refactor-lintrunner-config-to-use-uv-run 2025-12-04T08:53:58.7284863Z * [new branch] compatiblpy39util -> origin/compatiblpy39util 2025-12-04T08:53:58.7285044Z * [new branch] cond_hop_device -> origin/cond_hop_device 2025-12-04T08:53:58.7285221Z * [new branch] context_test -> origin/context_test 2025-12-04T08:53:58.7285468Z * [new branch] copilot/code-style-cleanup-python-pip -> origin/copilot/code-style-cleanup-python-pip 2025-12-04T08:53:58.7285721Z * [new branch] cpio/fix_new_ami_tests -> origin/cpio/fix_new_ami_tests 2025-12-04T08:53:58.7285948Z * [new branch] cpp-docs-dependency-upgrade -> origin/cpp-docs-dependency-upgrade 2025-12-04T08:53:58.7286173Z * [new branch] csl/always_produce_xml -> origin/csl/always_produce_xml 2025-12-04T08:53:58.7286377Z * [new branch] csl/build_test_more_procs -> origin/csl/build_test_more_procs 2025-12-04T08:53:58.7286595Z * [new branch] csl/build_test_more_procs2 -> origin/csl/build_test_more_procs2 2025-12-04T08:53:58.7286851Z * [new branch] csl/clean_up -> origin/csl/clean_up 2025-12-04T08:53:58.7287047Z * [new branch] csl/fix_retry_segfault_exit -> origin/csl/fix_retry_segfault_exit 2025-12-04T08:53:58.7287242Z * [new branch] csl/katex -> origin/csl/katex 2025-12-04T08:53:58.7287420Z * [new branch] csl/larger_runner -> origin/csl/larger_runner 2025-12-04T08:53:58.7287603Z * [new branch] csl/lint_testing -> origin/csl/lint_testing 2025-12-04T08:53:58.7287784Z * [new branch] csl/lint_thing -> origin/csl/lint_thing 2025-12-04T08:53:58.7287969Z * [new branch] csl/lintrunner_stuff -> origin/csl/lintrunner_stuff 2025-12-04T08:53:58.7288171Z * [new branch] csl/manually_gen_json -> origin/csl/manually_gen_json 2025-12-04T08:53:58.7288366Z * [new branch] csl/mps_sharding -> origin/csl/mps_sharding 2025-12-04T08:53:58.7288554Z * [new branch] csl/multistage_docker -> origin/csl/multistage_docker 2025-12-04T08:53:58.7288745Z * [new branch] csl/print_timing -> origin/csl/print_timing 2025-12-04T08:53:58.7288934Z * [new branch] csl/remove_experiment -> origin/csl/remove_experiment 2025-12-04T08:53:58.7289186Z * [new branch] csl/remove_maybe_unused_var -> origin/csl/remove_maybe_unused_var 2025-12-04T08:53:58.7289426Z * [new branch] csl/remove_repo_specific_autolabel -> origin/csl/remove_repo_specific_autolabel 2025-12-04T08:53:58.7289659Z * [new branch] csl/remove_run_parallel -> origin/csl/remove_run_parallel 2025-12-04T08:53:58.7289857Z * [new branch] csl/remove_unused_vars -> origin/csl/remove_unused_vars 2025-12-04T08:53:58.7290053Z * [new branch] csl/revert_open -> origin/csl/revert_open 2025-12-04T08:53:58.7290232Z * [new branch] csl/skip_build -> origin/csl/skip_build 2025-12-04T08:53:58.7290428Z * [new branch] csl/smaller_avx_amx_runenrs -> origin/csl/smaller_avx_amx_runenrs 2025-12-04T08:53:58.7290637Z * [new branch] csl/td_job_level -> origin/csl/td_job_level 2025-12-04T08:53:58.7290851Z * [new branch] csl/test_cuda_build_large_runner -> origin/csl/test_cuda_build_large_runner 2025-12-04T08:53:58.7291100Z * [new branch] csl/test_owners_autograd_dispatch_nn -> origin/csl/test_owners_autograd_dispatch_nn 2025-12-04T08:53:58.7291352Z * [new branch] csl/test_owners_higher_confidence -> origin/csl/test_owners_higher_confidence 2025-12-04T08:53:58.7291627Z * [new branch] csl/upload_json_running -> origin/csl/upload_json_running 2025-12-04T08:53:58.7291819Z * [new branch] csl/win_sccache -> origin/csl/win_sccache 2025-12-04T08:53:58.7291988Z * [new branch] csl/xml_stuff -> origin/csl/xml_stuff 2025-12-04T08:53:58.7292161Z * [new branch] cublasrelax2 -> origin/cublasrelax2 2025-12-04T08:53:58.7292329Z * [new branch] cuda_mempool -> origin/cuda_mempool 2025-12-04T08:53:58.7292508Z * [new branch] custom_lowering_dict -> origin/custom_lowering_dict 2025-12-04T08:53:58.7292716Z * [new branch] d4l3k/debug_plane_frtrace -> origin/d4l3k/debug_plane_frtrace 2025-12-04T08:53:58.7292903Z * [new branch] daxia6/2.8o3 -> origin/daxia6/2.8o3 2025-12-04T08:53:58.7293073Z * [new branch] debug-guard -> origin/debug-guard 2025-12-04T08:53:58.7293258Z * [new branch] delete-quant-docs -> origin/delete-quant-docs 2025-12-04T08:53:58.7293591Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.0 2025-12-04T08:53:58.7294044Z * [new branch] dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 -> origin/dependabot/pip/dot-ci/docker/ci_commit_pins/main/transformers-4.57.1 2025-12-04T08:53:58.7294379Z * [new branch] desertfire/test_cpp_wrapper -> origin/desertfire/test_cpp_wrapper 2025-12-04T08:53:58.7294628Z * [new branch] desertfire/triton-cpu-for-aarch64 -> origin/desertfire/triton-cpu-for-aarch64 2025-12-04T08:53:58.7294862Z * [new branch] dev/dhruva/flex_attn_opt -> origin/dev/dhruva/flex_attn_opt 2025-12-04T08:53:58.7295066Z * [new branch] dev/joona/MPSNDArrayAdd -> origin/dev/joona/MPSNDArrayAdd 2025-12-04T08:53:58.7295264Z * [new branch] dev/joona/Unranked -> origin/dev/joona/Unranked 2025-12-04T08:53:58.7295448Z * [new branch] dev/joona/cat -> origin/dev/joona/cat 2025-12-04T08:53:58.7295642Z * [new branch] dev/joona/embeddingbag -> origin/dev/joona/embeddingbag 2025-12-04T08:53:58.7295850Z * [new branch] dev/joona/fix_sdpa_memtest -> origin/dev/joona/fix_sdpa_memtest 2025-12-04T08:53:58.7296072Z * [new branch] dev/joona/getTensorsString -> origin/dev/joona/getTensorsString 2025-12-04T08:53:58.7296299Z * [new branch] dev/joona/mps_linear_macos14 -> origin/dev/joona/mps_linear_macos14 2025-12-04T08:53:58.7296531Z * [new branch] dev/joona/scalar_clamp -> origin/dev/joona/scalar_clamp 2025-12-04T08:53:58.7296722Z * [new branch] dev/joona/sdpa -> origin/dev/joona/sdpa 2025-12-04T08:53:58.7296956Z * [new branch] dev/joona/sdpa_api -> origin/dev/joona/sdpa_api 2025-12-04T08:53:58.7297143Z * [new branch] dev/joona/type_inf -> origin/dev/joona/type_inf 2025-12-04T08:53:58.7297350Z * [new branch] dev/joona/ulpAssertClose -> origin/dev/joona/ulpAssertClose 2025-12-04T08:53:58.7297553Z * [new branch] dev/joona/upsize3d -> origin/dev/joona/upsize3d 2025-12-04T08:53:58.7297731Z * [new branch] disp_counter -> origin/disp_counter 2025-12-04T08:53:58.7297917Z * [new branch] divyanshk-patch-1 -> origin/divyanshk-patch-1 2025-12-04T08:53:58.7298095Z * [new branch] docs -> origin/docs 2025-12-04T08:53:58.7298269Z * [new branch] documentation -> origin/documentation 2025-12-04T08:53:58.7298458Z * [new branch] eager_model_benchmarks -> origin/eager_model_benchmarks 2025-12-04T08:53:58.7298674Z * [new branch] embg/test_inductor_ci_control -> origin/embg/test_inductor_ci_control 2025-12-04T08:53:58.7298940Z * [new branch] embg/triton_l2_prefetch_128B -> origin/embg/triton_l2_prefetch_128B 2025-12-04T08:53:58.7299167Z * [new branch] embg/triton_l2_prefetch_256B -> origin/embg/triton_l2_prefetch_256B 2025-12-04T08:53:58.7299368Z * [new branch] eqy-patch-1 -> origin/eqy-patch-1 2025-12-04T08:53:58.7299539Z * [new branch] eqy-patch-2 -> origin/eqy-patch-2 2025-12-04T08:53:58.7299712Z * [new branch] eqy-patch-3 -> origin/eqy-patch-3 2025-12-04T08:53:58.7299879Z * [new branch] eqy-patch-4 -> origin/eqy-patch-4 2025-12-04T08:53:58.7300061Z * [new branch] eqy-patch-5 -> origin/eqy-patch-5 2025-12-04T08:53:58.7300231Z * [new branch] eqy-patch-6 -> origin/eqy-patch-6 2025-12-04T08:53:58.7300415Z * [new branch] exclamaforte/amd-ma -> origin/exclamaforte/amd-ma 2025-12-04T08:53:58.7300663Z * [new branch] exclamaforte/combo-kernels-perf-run -> origin/exclamaforte/combo-kernels-perf-run 2025-12-04T08:53:58.7300930Z * [new branch] exclamaforte/do_bench_refactor -> origin/exclamaforte/do_bench_refactor 2025-12-04T08:53:58.7301184Z * [new branch] exclamaforte/enable-mem-dep-fusion -> origin/exclamaforte/enable-mem-dep-fusion 2025-12-04T08:53:58.7301474Z * [new branch] exclamaforte/fix-exhaustive-autotuning -> origin/exclamaforte/fix-exhaustive-autotuning 2025-12-04T08:53:58.7301771Z * [new branch] exclamaforte/fix-trace-parsing-fx-svg -> origin/exclamaforte/fix-trace-parsing-fx-svg 2025-12-04T08:53:58.7302083Z * [new branch] exclamaforte/force-pointwise-cat-perf-run -> origin/exclamaforte/force-pointwise-cat-perf-run 2025-12-04T08:53:58.7302352Z * [new branch] exclamaforte/fusion-data -> origin/exclamaforte/fusion-data 2025-12-04T08:53:58.7302587Z * [new branch] exclamaforte/gemm-benchmark-run -> origin/exclamaforte/gemm-benchmark-run 2025-12-04T08:53:58.7302843Z * [new branch] exclamaforte/gemm-export-model -> origin/exclamaforte/gemm-export-model 2025-12-04T08:53:58.7303071Z * [new branch] exclamaforte/gemm-model -> origin/exclamaforte/gemm-model 2025-12-04T08:53:58.7303344Z * [new branch] exclamaforte/gemm-model-all-data-collection -> origin/exclamaforte/gemm-model-all-data-collection 2025-12-04T08:53:58.7303618Z * [new branch] exclamaforte/gemm-to-amd -> origin/exclamaforte/gemm-to-amd 2025-12-04T08:53:58.7303876Z * [new branch] exclamaforte/just-gemm-model -> origin/exclamaforte/just-gemm-model 2025-12-04T08:53:58.7304150Z * [new branch] exclamaforte/just-gemm-model-no-refactor -> origin/exclamaforte/just-gemm-model-no-refactor 2025-12-04T08:53:58.7304427Z * [new branch] exclamaforte/profile-diff-algo -> origin/exclamaforte/profile-diff-algo 2025-12-04T08:53:58.7304701Z * [new branch] exclamaforte/profiler-visualization -> origin/exclamaforte/profiler-visualization 2025-12-04T08:53:58.7304972Z * [new branch] exclamaforte/test_cpp_wrapper_mode -> origin/exclamaforte/test_cpp_wrapper_mode 2025-12-04T08:53:58.7305240Z * [new branch] exclamaforte/update-autotune-configs -> origin/exclamaforte/update-autotune-configs 2025-12-04T08:53:58.7305540Z * [new branch] exclamaforte/update-autotune-configs-2 -> origin/exclamaforte/update-autotune-configs-2 2025-12-04T08:53:58.7305770Z * [new branch] exec -> origin/exec 2025-12-04T08:53:58.7305949Z * [new branch] experimental-mosaic -> origin/experimental-mosaic 2025-12-04T08:53:58.7306142Z * [new branch] export-D61047529 -> origin/export-D61047529 2025-12-04T08:53:58.7306326Z * [new branch] export-D71412006 -> origin/export-D71412006 2025-12-04T08:53:58.7306525Z * [new branch] export-D73042989 -> origin/export-D73042989 2025-12-04T08:53:58.7306821Z * [new branch] export-D78957093 -> origin/export-D78957093 2025-12-04T08:53:58.7307008Z * [new branch] export-D78996107 -> origin/export-D78996107 2025-12-04T08:53:58.7307183Z * [new branch] export-D80823877 -> origin/export-D80823877 2025-12-04T08:53:58.7307355Z * [new branch] export-D80958642 -> origin/export-D80958642 2025-12-04T08:53:58.7307529Z * [new branch] export-D81054193 -> origin/export-D81054193 2025-12-04T08:53:58.7307706Z * [new branch] export-D81204584 -> origin/export-D81204584 2025-12-04T08:53:58.7307883Z * [new branch] export-D81429090 -> origin/export-D81429090 2025-12-04T08:53:58.7308057Z * [new branch] export-D82250826 -> origin/export-D82250826 2025-12-04T08:53:58.7308230Z * [new branch] export-D82253817 -> origin/export-D82253817 2025-12-04T08:53:58.7308405Z * [new branch] export-D83541846 -> origin/export-D83541846 2025-12-04T08:53:58.7308583Z * [new branch] export-D83627170 -> origin/export-D83627170 2025-12-04T08:53:58.7308761Z * [new branch] export-D83766701 -> origin/export-D83766701 2025-12-04T08:53:58.7308933Z * [new branch] export-D83768878 -> origin/export-D83768878 2025-12-04T08:53:58.7309106Z * [new branch] export-D83769447 -> origin/export-D83769447 2025-12-04T08:53:58.7309282Z * [new branch] export-D84089824 -> origin/export-D84089824 2025-12-04T08:53:58.7309454Z * [new branch] export-D84213020 -> origin/export-D84213020 2025-12-04T08:53:58.7309625Z * [new branch] export-D84373821 -> origin/export-D84373821 2025-12-04T08:53:58.7309797Z * [new branch] export-D84612194 -> origin/export-D84612194 2025-12-04T08:53:58.7309975Z * [new branch] export-D84890985 -> origin/export-D84890985 2025-12-04T08:53:58.7310147Z * [new branch] export-D85122326 -> origin/export-D85122326 2025-12-04T08:53:58.7310320Z * [new branch] export-D86256198 -> origin/export-D86256198 2025-12-04T08:53:58.7310497Z * [new branch] export-D86460608 -> origin/export-D86460608 2025-12-04T08:53:58.7310677Z * [new branch] export-D86474796 -> origin/export-D86474796 2025-12-04T08:53:58.7310849Z * [new branch] export-D86712396 -> origin/export-D86712396 2025-12-04T08:53:58.7311064Z * [new branch] export-D87022129 -> origin/export-D87022129 2025-12-04T08:53:58.7311235Z * [new branch] export-D87838959 -> origin/export-D87838959 2025-12-04T08:53:58.7311413Z * [new branch] export-D88319437 -> origin/export-D88319437 2025-12-04T08:53:58.7311636Z * [new branch] exported-model-train-idempotent -> origin/exported-model-train-idempotent 2025-12-04T08:53:58.7311872Z * [new branch] ezyang-titan-october -> origin/ezyang-titan-october 2025-12-04T08:53:58.7312071Z * [new branch] ezyang-titan-october2 -> origin/ezyang-titan-october2 2025-12-04T08:53:58.7312256Z * [new branch] ezyang-war -> origin/ezyang-war 2025-12-04T08:53:58.7312454Z * [new branch] ezyang/wip-aot-descriptors -> origin/ezyang/wip-aot-descriptors 2025-12-04T08:53:58.7312654Z * [new branch] fa_u8_brgemm -> origin/fa_u8_brgemm 2025-12-04T08:53:58.7312847Z * [new branch] fadeputr/sequence_fbgemm -> origin/fadeputr/sequence_fbgemm 2025-12-04T08:53:58.7313044Z * [new branch] fastmath_baseline -> origin/fastmath_baseline 2025-12-04T08:53:58.7313219Z * [new branch] fbcode/warm -> origin/fbcode/warm 2025-12-04T08:53:58.7313415Z * [new branch] fca -> origin/fca 2025-12-04T08:53:58.7313584Z * [new branch] fca2_ca5984c -> origin/fca2_ca5984c 2025-12-04T08:53:58.7313746Z * [new branch] fca5 -> origin/fca5 2025-12-04T08:53:58.7313921Z * [new branch] feature/justknobs-cpp -> origin/feature/justknobs-cpp 2025-12-04T08:53:58.7314121Z * [new branch] feature/numa-forkserver -> origin/feature/numa-forkserver 2025-12-04T08:53:58.7314315Z * [new branch] ffast_math_baseline -> origin/ffast_math_baseline 2025-12-04T08:53:58.7314505Z * [new branch] ffast_math_target -> origin/ffast_math_target 2025-12-04T08:53:58.7314685Z * [new branch] findhao/base_commit -> origin/findhao/base_commit 2025-12-04T08:53:58.7314876Z * [new branch] findhao/base_commit1 -> origin/findhao/base_commit1 2025-12-04T08:53:58.7315070Z * [new branch] findhao/multistream2 -> origin/findhao/multistream2 2025-12-04T08:53:58.7315266Z * [new branch] findhao/multistream5 -> origin/findhao/multistream5 2025-12-04T08:53:58.7315454Z * [new branch] findhao/multistream6 -> origin/findhao/multistream6 2025-12-04T08:53:58.7315649Z * [new branch] findhao/operatorbench3 -> origin/findhao/operatorbench3 2025-12-04T08:53:58.7315851Z * [new branch] findhao/operatorbench5 -> origin/findhao/operatorbench5 2025-12-04T08:53:58.7316043Z * [new branch] findhao/tritonparse -> origin/findhao/tritonparse 2025-12-04T08:53:58.7316263Z * [new branch] fix-ck-gemm-template-format -> origin/fix-ck-gemm-template-format 2025-12-04T08:53:58.7316473Z * [new branch] fix-config-ignore -> origin/fix-config-ignore 2025-12-04T08:53:58.7316656Z * [new branch] fix-dict-guard -> origin/fix-dict-guard 2025-12-04T08:53:58.7316884Z * [new branch] fix_addmm_issue -> origin/fix_addmm_issue 2025-12-04T08:53:58.7317086Z * [new branch] fix_amd_missing_cluster_dims -> origin/fix_amd_missing_cluster_dims 2025-12-04T08:53:58.7317285Z * [new branch] fix_bench_bwd_pass -> origin/fix_bench_bwd_pass 2025-12-04T08:53:58.7317481Z * [new branch] fix_mem_profiler_config -> origin/fix_mem_profiler_config 2025-12-04T08:53:58.7317673Z * [new branch] fix_nvrtc_discovery -> origin/fix_nvrtc_discovery 2025-12-04T08:53:58.7317848Z * [new branch] fix_op_runner -> origin/fix_op_runner 2025-12-04T08:53:58.7318056Z * [new branch] fix_ubn_159469 -> origin/fix_ubn_159469 2025-12-04T08:53:58.7318230Z * [new branch] fixes-triage -> origin/fixes-triage 2025-12-04T08:53:58.7318406Z * [new branch] fixflashinfer -> origin/fixflashinfer 2025-12-04T08:53:58.7318588Z * [new branch] flash_decoding_cpu -> origin/flash_decoding_cpu 2025-12-04T08:53:58.7318766Z * [new branch] flex-flash -> origin/flex-flash 2025-12-04T08:53:58.7318963Z * [new branch] flex_attention_functorch_grad -> origin/flex_attention_functorch_grad 2025-12-04T08:53:58.7319169Z * [new branch] flex_flash -> origin/flex_flash 2025-12-04T08:53:58.7319372Z * [new branch] fmassa/fix_memeff_sharding_rule -> origin/fmassa/fix_memeff_sharding_rule 2025-12-04T08:53:58.7319619Z * [new branch] fmassa/tests_comm_compute_scheduler -> origin/fmassa/tests_comm_compute_scheduler 2025-12-04T08:53:58.7319838Z * [new branch] forkserver_fix -> origin/forkserver_fix 2025-12-04T08:53:58.7320016Z * [new branch] fsdp2_trace_rules -> origin/fsdp2_trace_rules 2025-12-04T08:53:58.7320186Z * [new branch] fx_cpp -> origin/fx_cpp 2025-12-04T08:53:58.7320392Z * [new branch] fy/fix-win -> origin/fy/fix-win 2025-12-04T08:53:58.7320561Z * [new branch] galv-patch-1 -> origin/galv-patch-1 2025-12-04T08:53:58.7320788Z * [new branch] galv/cudagraphs-conditional-nodes-4 -> origin/galv/cudagraphs-conditional-nodes-4 2025-12-04T08:53:58.7321053Z * [new branch] georgehong/cmakelists-patch -> origin/georgehong/cmakelists-patch 2025-12-04T08:53:58.7321263Z * [new branch] gh/AlnisM/1/base -> origin/gh/AlnisM/1/base 2025-12-04T08:53:58.7321443Z * [new branch] gh/AlnisM/1/head -> origin/gh/AlnisM/1/head 2025-12-04T08:53:58.7321629Z * [new branch] gh/EikanWang/67/base -> origin/gh/EikanWang/67/base 2025-12-04T08:53:58.7321817Z * [new branch] gh/EikanWang/67/head -> origin/gh/EikanWang/67/head 2025-12-04T08:53:58.7322013Z * [new branch] gh/Gasoonjia/1/base -> origin/gh/Gasoonjia/1/base 2025-12-04T08:53:58.7322203Z * [new branch] gh/Gasoonjia/1/head -> origin/gh/Gasoonjia/1/head 2025-12-04T08:53:58.7322386Z * [new branch] gh/H-Huang/131/base -> origin/gh/H-Huang/131/base 2025-12-04T08:53:58.7322573Z * [new branch] gh/H-Huang/131/head -> origin/gh/H-Huang/131/head 2025-12-04T08:53:58.7322751Z * [new branch] gh/H-Huang/131/orig -> origin/gh/H-Huang/131/orig 2025-12-04T08:53:58.7322927Z * [new branch] gh/H-Huang/132/base -> origin/gh/H-Huang/132/base 2025-12-04T08:53:58.7323110Z * [new branch] gh/H-Huang/132/head -> origin/gh/H-Huang/132/head 2025-12-04T08:53:58.7323291Z * [new branch] gh/H-Huang/132/orig -> origin/gh/H-Huang/132/orig 2025-12-04T08:53:58.7323471Z * [new branch] gh/H-Huang/180/base -> origin/gh/H-Huang/180/base 2025-12-04T08:53:58.7323649Z * [new branch] gh/H-Huang/180/head -> origin/gh/H-Huang/180/head 2025-12-04T08:53:58.7323827Z * [new branch] gh/H-Huang/180/orig -> origin/gh/H-Huang/180/orig 2025-12-04T08:53:58.7324003Z * [new branch] gh/H-Huang/182/base -> origin/gh/H-Huang/182/base 2025-12-04T08:53:58.7324182Z * [new branch] gh/H-Huang/182/head -> origin/gh/H-Huang/182/head 2025-12-04T08:53:58.7324478Z * [new branch] gh/H-Huang/182/orig -> origin/gh/H-Huang/182/orig 2025-12-04T08:53:58.7324656Z * [new branch] gh/H-Huang/226/base -> origin/gh/H-Huang/226/base 2025-12-04T08:53:58.7324866Z * [new branch] gh/H-Huang/226/head -> origin/gh/H-Huang/226/head 2025-12-04T08:53:58.7325045Z * [new branch] gh/H-Huang/226/orig -> origin/gh/H-Huang/226/orig 2025-12-04T08:53:58.7325226Z * [new branch] gh/H-Huang/228/base -> origin/gh/H-Huang/228/base 2025-12-04T08:53:58.7325404Z * [new branch] gh/H-Huang/228/head -> origin/gh/H-Huang/228/head 2025-12-04T08:53:58.7325580Z * [new branch] gh/H-Huang/228/orig -> origin/gh/H-Huang/228/orig 2025-12-04T08:53:58.7325774Z * [new branch] gh/IvanKobzarev/150/base -> origin/gh/IvanKobzarev/150/base 2025-12-04T08:53:58.7325980Z * [new branch] gh/IvanKobzarev/150/head -> origin/gh/IvanKobzarev/150/head 2025-12-04T08:53:58.7326184Z * [new branch] gh/IvanKobzarev/150/orig -> origin/gh/IvanKobzarev/150/orig 2025-12-04T08:53:58.7326391Z * [new branch] gh/IvanKobzarev/157/base -> origin/gh/IvanKobzarev/157/base 2025-12-04T08:53:58.7326602Z * [new branch] gh/IvanKobzarev/157/head -> origin/gh/IvanKobzarev/157/head 2025-12-04T08:53:58.7326842Z * [new branch] gh/IvanKobzarev/157/orig -> origin/gh/IvanKobzarev/157/orig 2025-12-04T08:53:58.7327053Z * [new branch] gh/IvanKobzarev/159/base -> origin/gh/IvanKobzarev/159/base 2025-12-04T08:53:58.7327300Z * [new branch] gh/IvanKobzarev/159/head -> origin/gh/IvanKobzarev/159/head 2025-12-04T08:53:58.7327501Z * [new branch] gh/IvanKobzarev/159/orig -> origin/gh/IvanKobzarev/159/orig 2025-12-04T08:53:58.7327710Z * [new branch] gh/IvanKobzarev/162/base -> origin/gh/IvanKobzarev/162/base 2025-12-04T08:53:58.7327920Z * [new branch] gh/IvanKobzarev/162/head -> origin/gh/IvanKobzarev/162/head 2025-12-04T08:53:58.7328125Z * [new branch] gh/IvanKobzarev/162/orig -> origin/gh/IvanKobzarev/162/orig 2025-12-04T08:53:58.7328331Z * [new branch] gh/IvanKobzarev/163/base -> origin/gh/IvanKobzarev/163/base 2025-12-04T08:53:58.7328545Z * [new branch] gh/IvanKobzarev/163/head -> origin/gh/IvanKobzarev/163/head 2025-12-04T08:53:58.7328747Z * [new branch] gh/IvanKobzarev/163/orig -> origin/gh/IvanKobzarev/163/orig 2025-12-04T08:53:58.7328953Z * [new branch] gh/IvanKobzarev/166/base -> origin/gh/IvanKobzarev/166/base 2025-12-04T08:53:58.7329163Z * [new branch] gh/IvanKobzarev/166/head -> origin/gh/IvanKobzarev/166/head 2025-12-04T08:53:58.7329367Z * [new branch] gh/IvanKobzarev/166/orig -> origin/gh/IvanKobzarev/166/orig 2025-12-04T08:53:58.7329573Z * [new branch] gh/IvanKobzarev/167/base -> origin/gh/IvanKobzarev/167/base 2025-12-04T08:53:58.7329773Z * [new branch] gh/IvanKobzarev/167/head -> origin/gh/IvanKobzarev/167/head 2025-12-04T08:53:58.7329972Z * [new branch] gh/IvanKobzarev/167/orig -> origin/gh/IvanKobzarev/167/orig 2025-12-04T08:53:58.7330184Z * [new branch] gh/IvanKobzarev/168/base -> origin/gh/IvanKobzarev/168/base 2025-12-04T08:53:58.7330390Z * [new branch] gh/IvanKobzarev/168/head -> origin/gh/IvanKobzarev/168/head 2025-12-04T08:53:58.7330590Z * [new branch] gh/IvanKobzarev/168/orig -> origin/gh/IvanKobzarev/168/orig 2025-12-04T08:53:58.7330801Z * [new branch] gh/IvanKobzarev/169/base -> origin/gh/IvanKobzarev/169/base 2025-12-04T08:53:58.7331015Z * [new branch] gh/IvanKobzarev/169/head -> origin/gh/IvanKobzarev/169/head 2025-12-04T08:53:58.7331220Z * [new branch] gh/IvanKobzarev/169/orig -> origin/gh/IvanKobzarev/169/orig 2025-12-04T08:53:58.7331427Z * [new branch] gh/IvanKobzarev/170/base -> origin/gh/IvanKobzarev/170/base 2025-12-04T08:53:58.7331635Z * [new branch] gh/IvanKobzarev/170/head -> origin/gh/IvanKobzarev/170/head 2025-12-04T08:53:58.7331835Z * [new branch] gh/IvanKobzarev/170/orig -> origin/gh/IvanKobzarev/170/orig 2025-12-04T08:53:58.7332070Z * [new branch] gh/IvanKobzarev/171/base -> origin/gh/IvanKobzarev/171/base 2025-12-04T08:53:58.7332270Z * [new branch] gh/IvanKobzarev/171/head -> origin/gh/IvanKobzarev/171/head 2025-12-04T08:53:58.7332478Z * [new branch] gh/IvanKobzarev/171/orig -> origin/gh/IvanKobzarev/171/orig 2025-12-04T08:53:58.7332683Z * [new branch] gh/IvanKobzarev/172/base -> origin/gh/IvanKobzarev/172/base 2025-12-04T08:53:58.7332884Z * [new branch] gh/IvanKobzarev/172/head -> origin/gh/IvanKobzarev/172/head 2025-12-04T08:53:58.7333087Z * [new branch] gh/IvanKobzarev/172/orig -> origin/gh/IvanKobzarev/172/orig 2025-12-04T08:53:58.7333296Z * [new branch] gh/IvanKobzarev/173/base -> origin/gh/IvanKobzarev/173/base 2025-12-04T08:53:58.7333498Z * [new branch] gh/IvanKobzarev/173/head -> origin/gh/IvanKobzarev/173/head 2025-12-04T08:53:58.7333707Z * [new branch] gh/IvanKobzarev/173/orig -> origin/gh/IvanKobzarev/173/orig 2025-12-04T08:53:58.7333910Z * [new branch] gh/IvanKobzarev/174/base -> origin/gh/IvanKobzarev/174/base 2025-12-04T08:53:58.7334112Z * [new branch] gh/IvanKobzarev/174/head -> origin/gh/IvanKobzarev/174/head 2025-12-04T08:53:58.7334343Z * [new branch] gh/IvanKobzarev/174/orig -> origin/gh/IvanKobzarev/174/orig 2025-12-04T08:53:58.7334548Z * [new branch] gh/IvanKobzarev/175/base -> origin/gh/IvanKobzarev/175/base 2025-12-04T08:53:58.7334746Z * [new branch] gh/IvanKobzarev/175/head -> origin/gh/IvanKobzarev/175/head 2025-12-04T08:53:58.7334951Z * [new branch] gh/IvanKobzarev/175/orig -> origin/gh/IvanKobzarev/175/orig 2025-12-04T08:53:58.7335159Z * [new branch] gh/IvanKobzarev/176/base -> origin/gh/IvanKobzarev/176/base 2025-12-04T08:53:58.7335359Z * [new branch] gh/IvanKobzarev/176/head -> origin/gh/IvanKobzarev/176/head 2025-12-04T08:53:58.7335565Z * [new branch] gh/IvanKobzarev/176/orig -> origin/gh/IvanKobzarev/176/orig 2025-12-04T08:53:58.7335775Z * [new branch] gh/IvanKobzarev/177/base -> origin/gh/IvanKobzarev/177/base 2025-12-04T08:53:58.7335978Z * [new branch] gh/IvanKobzarev/177/head -> origin/gh/IvanKobzarev/177/head 2025-12-04T08:53:58.7336187Z * [new branch] gh/IvanKobzarev/177/orig -> origin/gh/IvanKobzarev/177/orig 2025-12-04T08:53:58.7336393Z * [new branch] gh/IvanKobzarev/178/base -> origin/gh/IvanKobzarev/178/base 2025-12-04T08:53:58.7336593Z * [new branch] gh/IvanKobzarev/178/head -> origin/gh/IvanKobzarev/178/head 2025-12-04T08:53:58.7336841Z * [new branch] gh/IvanKobzarev/178/orig -> origin/gh/IvanKobzarev/178/orig 2025-12-04T08:53:58.7337048Z * [new branch] gh/IvanKobzarev/179/base -> origin/gh/IvanKobzarev/179/base 2025-12-04T08:53:58.7337251Z * [new branch] gh/IvanKobzarev/179/head -> origin/gh/IvanKobzarev/179/head 2025-12-04T08:53:58.7337454Z * [new branch] gh/IvanKobzarev/179/orig -> origin/gh/IvanKobzarev/179/orig 2025-12-04T08:53:58.7337657Z * [new branch] gh/IvanKobzarev/180/base -> origin/gh/IvanKobzarev/180/base 2025-12-04T08:53:58.7337865Z * [new branch] gh/IvanKobzarev/180/head -> origin/gh/IvanKobzarev/180/head 2025-12-04T08:53:58.7338070Z * [new branch] gh/IvanKobzarev/180/orig -> origin/gh/IvanKobzarev/180/orig 2025-12-04T08:53:58.7338277Z * [new branch] gh/IvanKobzarev/181/base -> origin/gh/IvanKobzarev/181/base 2025-12-04T08:53:58.7338476Z * [new branch] gh/IvanKobzarev/181/head -> origin/gh/IvanKobzarev/181/head 2025-12-04T08:53:58.7338685Z * [new branch] gh/IvanKobzarev/181/orig -> origin/gh/IvanKobzarev/181/orig 2025-12-04T08:53:58.7338889Z * [new branch] gh/IvanKobzarev/182/base -> origin/gh/IvanKobzarev/182/base 2025-12-04T08:53:58.7339139Z * [new branch] gh/IvanKobzarev/182/head -> origin/gh/IvanKobzarev/182/head 2025-12-04T08:53:58.7339346Z * [new branch] gh/IvanKobzarev/182/orig -> origin/gh/IvanKobzarev/182/orig 2025-12-04T08:53:58.7339548Z * [new branch] gh/IvanKobzarev/183/base -> origin/gh/IvanKobzarev/183/base 2025-12-04T08:53:58.7339753Z * [new branch] gh/IvanKobzarev/183/head -> origin/gh/IvanKobzarev/183/head 2025-12-04T08:53:58.7339955Z * [new branch] gh/IvanKobzarev/183/orig -> origin/gh/IvanKobzarev/183/orig 2025-12-04T08:53:58.7340156Z * [new branch] gh/IvanKobzarev/184/base -> origin/gh/IvanKobzarev/184/base 2025-12-04T08:53:58.7340359Z * [new branch] gh/IvanKobzarev/184/head -> origin/gh/IvanKobzarev/184/head 2025-12-04T08:53:58.7340569Z * [new branch] gh/IvanKobzarev/184/orig -> origin/gh/IvanKobzarev/184/orig 2025-12-04T08:53:58.7340774Z * [new branch] gh/NikhilAPatel/1/base -> origin/gh/NikhilAPatel/1/base 2025-12-04T08:53:58.7340981Z * [new branch] gh/NikhilAPatel/1/head -> origin/gh/NikhilAPatel/1/head 2025-12-04T08:53:58.7341183Z * [new branch] gh/NikhilAPatel/2/base -> origin/gh/NikhilAPatel/2/base 2025-12-04T08:53:58.7341387Z * [new branch] gh/NikhilAPatel/2/head -> origin/gh/NikhilAPatel/2/head 2025-12-04T08:53:58.7341619Z * [new branch] gh/NikhilAPatel/4/base -> origin/gh/NikhilAPatel/4/base 2025-12-04T08:53:58.7341819Z * [new branch] gh/NikhilAPatel/4/head -> origin/gh/NikhilAPatel/4/head 2025-12-04T08:53:58.7342014Z * [new branch] gh/NikhilAPatel/5/base -> origin/gh/NikhilAPatel/5/base 2025-12-04T08:53:58.7342218Z * [new branch] gh/NikhilAPatel/5/head -> origin/gh/NikhilAPatel/5/head 2025-12-04T08:53:58.7342415Z * [new branch] gh/NikhilAPatel/5/orig -> origin/gh/NikhilAPatel/5/orig 2025-12-04T08:53:58.7342607Z * [new branch] gh/PaliC/17/base -> origin/gh/PaliC/17/base 2025-12-04T08:53:58.7342789Z * [new branch] gh/PaliC/17/head -> origin/gh/PaliC/17/head 2025-12-04T08:53:58.7342971Z * [new branch] gh/PaliC/17/orig -> origin/gh/PaliC/17/orig 2025-12-04T08:53:58.7343151Z * [new branch] gh/PaliC/18/base -> origin/gh/PaliC/18/base 2025-12-04T08:53:58.7343332Z * [new branch] gh/PaliC/18/head -> origin/gh/PaliC/18/head 2025-12-04T08:53:58.7343511Z * [new branch] gh/PaliC/18/orig -> origin/gh/PaliC/18/orig 2025-12-04T08:53:58.7343685Z * [new branch] gh/PaliC/20/base -> origin/gh/PaliC/20/base 2025-12-04T08:53:58.7343861Z * [new branch] gh/PaliC/20/head -> origin/gh/PaliC/20/head 2025-12-04T08:53:58.7344042Z * [new branch] gh/PaliC/20/orig -> origin/gh/PaliC/20/orig 2025-12-04T08:53:58.7344223Z * [new branch] gh/PaliC/21/base -> origin/gh/PaliC/21/base 2025-12-04T08:53:58.7344404Z * [new branch] gh/PaliC/21/head -> origin/gh/PaliC/21/head 2025-12-04T08:53:58.7344579Z * [new branch] gh/PaliC/21/orig -> origin/gh/PaliC/21/orig 2025-12-04T08:53:58.7344760Z * [new branch] gh/PaliC/23/base -> origin/gh/PaliC/23/base 2025-12-04T08:53:58.7344947Z * [new branch] gh/PaliC/23/head -> origin/gh/PaliC/23/head 2025-12-04T08:53:58.7345123Z * [new branch] gh/PaliC/23/orig -> origin/gh/PaliC/23/orig 2025-12-04T08:53:58.7345301Z * [new branch] gh/PaliC/24/base -> origin/gh/PaliC/24/base 2025-12-04T08:53:58.7345482Z * [new branch] gh/PaliC/24/head -> origin/gh/PaliC/24/head 2025-12-04T08:53:58.7345656Z * [new branch] gh/PaliC/24/orig -> origin/gh/PaliC/24/orig 2025-12-04T08:53:58.7345871Z * [new branch] gh/PaliC/25/head -> origin/gh/PaliC/25/head 2025-12-04T08:53:58.7346050Z * [new branch] gh/PaliC/25/next -> origin/gh/PaliC/25/next 2025-12-04T08:53:58.7346225Z * [new branch] gh/PaliC/25/orig -> origin/gh/PaliC/25/orig 2025-12-04T08:53:58.7346404Z * [new branch] gh/PaliC/26/head -> origin/gh/PaliC/26/head 2025-12-04T08:53:58.7346586Z * [new branch] gh/PaliC/26/next -> origin/gh/PaliC/26/next 2025-12-04T08:53:58.7346815Z * [new branch] gh/PaliC/26/orig -> origin/gh/PaliC/26/orig 2025-12-04T08:53:58.7347275Z * [new branch] gh/PaliC/27/next -> origin/gh/PaliC/27/next 2025-12-04T08:53:58.7347459Z * [new branch] gh/PaliC/28/head -> origin/gh/PaliC/28/head 2025-12-04T08:53:58.7347635Z * [new branch] gh/PaliC/28/next -> origin/gh/PaliC/28/next 2025-12-04T08:53:58.7347814Z * [new branch] gh/PaliC/28/orig -> origin/gh/PaliC/28/orig 2025-12-04T08:53:58.7348000Z * [new branch] gh/PaliC/29/head -> origin/gh/PaliC/29/head 2025-12-04T08:53:58.7348174Z * [new branch] gh/PaliC/29/next -> origin/gh/PaliC/29/next 2025-12-04T08:53:58.7348365Z * [new branch] gh/PaliC/29/orig -> origin/gh/PaliC/29/orig 2025-12-04T08:53:58.7348578Z * [new branch] gh/PaliC/30/head -> origin/gh/PaliC/30/head 2025-12-04T08:53:58.7348757Z * [new branch] gh/PaliC/30/next -> origin/gh/PaliC/30/next 2025-12-04T08:53:58.7348936Z * [new branch] gh/PaliC/30/orig -> origin/gh/PaliC/30/orig 2025-12-04T08:53:58.7349110Z * [new branch] gh/PaliC/31/head -> origin/gh/PaliC/31/head 2025-12-04T08:53:58.7349294Z * [new branch] gh/PaliC/31/next -> origin/gh/PaliC/31/next 2025-12-04T08:53:58.7349473Z * [new branch] gh/PaliC/31/orig -> origin/gh/PaliC/31/orig 2025-12-04T08:53:58.7349664Z * [new branch] gh/PaulZhang12/25/base -> origin/gh/PaulZhang12/25/base 2025-12-04T08:53:58.7349868Z * [new branch] gh/PaulZhang12/25/head -> origin/gh/PaulZhang12/25/head 2025-12-04T08:53:58.7350068Z * [new branch] gh/PaulZhang12/25/orig -> origin/gh/PaulZhang12/25/orig 2025-12-04T08:53:58.7350266Z * [new branch] gh/PaulZhang12/28/base -> origin/gh/PaulZhang12/28/base 2025-12-04T08:53:58.7350464Z * [new branch] gh/PaulZhang12/28/head -> origin/gh/PaulZhang12/28/head 2025-12-04T08:53:58.7350660Z * [new branch] gh/PaulZhang12/28/orig -> origin/gh/PaulZhang12/28/orig 2025-12-04T08:53:58.7350852Z * [new branch] gh/PaulZhang12/31/base -> origin/gh/PaulZhang12/31/base 2025-12-04T08:53:58.7351050Z * [new branch] gh/PaulZhang12/31/head -> origin/gh/PaulZhang12/31/head 2025-12-04T08:53:58.7351244Z * [new branch] gh/PaulZhang12/31/orig -> origin/gh/PaulZhang12/31/orig 2025-12-04T08:53:58.7351437Z * [new branch] gh/PaulZhang12/37/base -> origin/gh/PaulZhang12/37/base 2025-12-04T08:53:58.7351631Z * [new branch] gh/PaulZhang12/37/head -> origin/gh/PaulZhang12/37/head 2025-12-04T08:53:58.7351826Z * [new branch] gh/PaulZhang12/37/orig -> origin/gh/PaulZhang12/37/orig 2025-12-04T08:53:58.7352019Z * [new branch] gh/PaulZhang12/40/base -> origin/gh/PaulZhang12/40/base 2025-12-04T08:53:58.7352210Z * [new branch] gh/PaulZhang12/40/head -> origin/gh/PaulZhang12/40/head 2025-12-04T08:53:58.7352399Z * [new branch] gh/PaulZhang12/40/orig -> origin/gh/PaulZhang12/40/orig 2025-12-04T08:53:58.7352587Z * [new branch] gh/PaulZhang12/42/base -> origin/gh/PaulZhang12/42/base 2025-12-04T08:53:58.7352777Z * [new branch] gh/PaulZhang12/42/head -> origin/gh/PaulZhang12/42/head 2025-12-04T08:53:58.7352998Z * [new branch] gh/PaulZhang12/43/base -> origin/gh/PaulZhang12/43/base 2025-12-04T08:53:58.7353188Z * [new branch] gh/PaulZhang12/43/head -> origin/gh/PaulZhang12/43/head 2025-12-04T08:53:58.7353378Z * [new branch] gh/PaulZhang12/43/orig -> origin/gh/PaulZhang12/43/orig 2025-12-04T08:53:58.7353566Z * [new branch] gh/PaulZhang12/44/base -> origin/gh/PaulZhang12/44/base 2025-12-04T08:53:58.7353760Z * [new branch] gh/PaulZhang12/44/head -> origin/gh/PaulZhang12/44/head 2025-12-04T08:53:58.7353950Z * [new branch] gh/PaulZhang12/45/base -> origin/gh/PaulZhang12/45/base 2025-12-04T08:53:58.7354139Z * [new branch] gh/PaulZhang12/45/head -> origin/gh/PaulZhang12/45/head 2025-12-04T08:53:58.7354329Z * [new branch] gh/PaulZhang12/45/orig -> origin/gh/PaulZhang12/45/orig 2025-12-04T08:53:58.7354519Z * [new branch] gh/PaulZhang12/46/base -> origin/gh/PaulZhang12/46/base 2025-12-04T08:53:58.7354710Z * [new branch] gh/PaulZhang12/46/head -> origin/gh/PaulZhang12/46/head 2025-12-04T08:53:58.7354900Z * [new branch] gh/PaulZhang12/46/orig -> origin/gh/PaulZhang12/46/orig 2025-12-04T08:53:58.7355090Z * [new branch] gh/PaulZhang12/47/base -> origin/gh/PaulZhang12/47/base 2025-12-04T08:53:58.7355278Z * [new branch] gh/PaulZhang12/47/head -> origin/gh/PaulZhang12/47/head 2025-12-04T08:53:58.7355498Z * [new branch] gh/PaulZhang12/47/orig -> origin/gh/PaulZhang12/47/orig 2025-12-04T08:53:58.7355692Z * [new branch] gh/PaulZhang12/48/base -> origin/gh/PaulZhang12/48/base 2025-12-04T08:53:58.7355882Z * [new branch] gh/PaulZhang12/48/head -> origin/gh/PaulZhang12/48/head 2025-12-04T08:53:58.7356076Z * [new branch] gh/PaulZhang12/48/orig -> origin/gh/PaulZhang12/48/orig 2025-12-04T08:53:58.7356269Z * [new branch] gh/SamGinzburg/11/base -> origin/gh/SamGinzburg/11/base 2025-12-04T08:53:58.7356462Z * [new branch] gh/SamGinzburg/11/head -> origin/gh/SamGinzburg/11/head 2025-12-04T08:53:58.7356662Z * [new branch] gh/SherlockNoMad/1/base -> origin/gh/SherlockNoMad/1/base 2025-12-04T08:53:58.7356915Z * [new branch] gh/SherlockNoMad/1/head -> origin/gh/SherlockNoMad/1/head 2025-12-04T08:53:58.7357121Z * [new branch] gh/SherlockNoMad/10/base -> origin/gh/SherlockNoMad/10/base 2025-12-04T08:53:58.7357323Z * [new branch] gh/SherlockNoMad/10/head -> origin/gh/SherlockNoMad/10/head 2025-12-04T08:53:58.7357522Z * [new branch] gh/SherlockNoMad/10/orig -> origin/gh/SherlockNoMad/10/orig 2025-12-04T08:53:58.7357726Z * [new branch] gh/SherlockNoMad/11/base -> origin/gh/SherlockNoMad/11/base 2025-12-04T08:53:58.7357928Z * [new branch] gh/SherlockNoMad/11/head -> origin/gh/SherlockNoMad/11/head 2025-12-04T08:53:58.7358131Z * [new branch] gh/SherlockNoMad/11/orig -> origin/gh/SherlockNoMad/11/orig 2025-12-04T08:53:58.7358333Z * [new branch] gh/SherlockNoMad/12/base -> origin/gh/SherlockNoMad/12/base 2025-12-04T08:53:58.7358538Z * [new branch] gh/SherlockNoMad/12/head -> origin/gh/SherlockNoMad/12/head 2025-12-04T08:53:58.7358737Z * [new branch] gh/SherlockNoMad/12/orig -> origin/gh/SherlockNoMad/12/orig 2025-12-04T08:53:58.7358957Z * [new branch] gh/SherlockNoMad/15/base -> origin/gh/SherlockNoMad/15/base 2025-12-04T08:53:58.7359172Z * [new branch] gh/SherlockNoMad/15/head -> origin/gh/SherlockNoMad/15/head 2025-12-04T08:53:58.7359369Z * [new branch] gh/SherlockNoMad/15/orig -> origin/gh/SherlockNoMad/15/orig 2025-12-04T08:53:58.7359572Z * [new branch] gh/SherlockNoMad/17/base -> origin/gh/SherlockNoMad/17/base 2025-12-04T08:53:58.7359782Z * [new branch] gh/SherlockNoMad/17/head -> origin/gh/SherlockNoMad/17/head 2025-12-04T08:53:58.7360017Z * [new branch] gh/SherlockNoMad/17/orig -> origin/gh/SherlockNoMad/17/orig 2025-12-04T08:53:58.7360217Z * [new branch] gh/SherlockNoMad/18/base -> origin/gh/SherlockNoMad/18/base 2025-12-04T08:53:58.7360417Z * [new branch] gh/SherlockNoMad/18/head -> origin/gh/SherlockNoMad/18/head 2025-12-04T08:53:58.7360615Z * [new branch] gh/SherlockNoMad/18/orig -> origin/gh/SherlockNoMad/18/orig 2025-12-04T08:53:58.7360815Z * [new branch] gh/SherlockNoMad/19/base -> origin/gh/SherlockNoMad/19/base 2025-12-04T08:53:58.7361015Z * [new branch] gh/SherlockNoMad/19/head -> origin/gh/SherlockNoMad/19/head 2025-12-04T08:53:58.7361215Z * [new branch] gh/SherlockNoMad/19/orig -> origin/gh/SherlockNoMad/19/orig 2025-12-04T08:53:58.7361415Z * [new branch] gh/SherlockNoMad/2/base -> origin/gh/SherlockNoMad/2/base 2025-12-04T08:53:58.7361618Z * [new branch] gh/SherlockNoMad/2/head -> origin/gh/SherlockNoMad/2/head 2025-12-04T08:53:58.7361822Z * [new branch] gh/SherlockNoMad/20/base -> origin/gh/SherlockNoMad/20/base 2025-12-04T08:53:58.7362027Z * [new branch] gh/SherlockNoMad/20/head -> origin/gh/SherlockNoMad/20/head 2025-12-04T08:53:58.7362234Z * [new branch] gh/SherlockNoMad/20/orig -> origin/gh/SherlockNoMad/20/orig 2025-12-04T08:53:58.7362464Z * [new branch] gh/SherlockNoMad/21/base -> origin/gh/SherlockNoMad/21/base 2025-12-04T08:53:58.7362670Z * [new branch] gh/SherlockNoMad/21/head -> origin/gh/SherlockNoMad/21/head 2025-12-04T08:53:58.7362878Z * [new branch] gh/SherlockNoMad/21/orig -> origin/gh/SherlockNoMad/21/orig 2025-12-04T08:53:58.7363077Z * [new branch] gh/SherlockNoMad/3/base -> origin/gh/SherlockNoMad/3/base 2025-12-04T08:53:58.7363282Z * [new branch] gh/SherlockNoMad/3/head -> origin/gh/SherlockNoMad/3/head 2025-12-04T08:53:58.7363490Z * [new branch] gh/SherlockNoMad/4/base -> origin/gh/SherlockNoMad/4/base 2025-12-04T08:53:58.7363687Z * [new branch] gh/SherlockNoMad/4/head -> origin/gh/SherlockNoMad/4/head 2025-12-04T08:53:58.7363889Z * [new branch] gh/SherlockNoMad/5/base -> origin/gh/SherlockNoMad/5/base 2025-12-04T08:53:58.7364110Z * [new branch] gh/SherlockNoMad/5/head -> origin/gh/SherlockNoMad/5/head 2025-12-04T08:53:58.7364355Z * [new branch] gh/Sidharth123-cpu/24/base -> origin/gh/Sidharth123-cpu/24/base 2025-12-04T08:53:58.7364577Z * [new branch] gh/Sidharth123-cpu/25/base -> origin/gh/Sidharth123-cpu/25/base 2025-12-04T08:53:58.7364791Z * [new branch] gh/Sidharth123-cpu/26/base -> origin/gh/Sidharth123-cpu/26/base 2025-12-04T08:53:58.7365008Z * [new branch] gh/Sidharth123-cpu/27/base -> origin/gh/Sidharth123-cpu/27/base 2025-12-04T08:53:58.7365215Z * [new branch] gh/StrongerXi/1/base -> origin/gh/StrongerXi/1/base 2025-12-04T08:53:58.7365414Z * [new branch] gh/StrongerXi/1/head -> origin/gh/StrongerXi/1/head 2025-12-04T08:53:58.7365621Z * [new branch] gh/StrongerXi/71/base -> origin/gh/StrongerXi/71/base 2025-12-04T08:53:58.7383559Z * [new branch] gh/StrongerXi/71/head -> origin/gh/StrongerXi/71/head 2025-12-04T08:53:58.7383830Z * [new branch] gh/StrongerXi/72/base -> origin/gh/StrongerXi/72/base 2025-12-04T08:53:58.7384063Z * [new branch] gh/StrongerXi/72/head -> origin/gh/StrongerXi/72/head 2025-12-04T08:53:58.7384281Z * [new branch] gh/StrongerXi/73/base -> origin/gh/StrongerXi/73/base 2025-12-04T08:53:58.7384475Z * [new branch] gh/StrongerXi/73/head -> origin/gh/StrongerXi/73/head 2025-12-04T08:53:58.7384670Z * [new branch] gh/StrongerXi/73/orig -> origin/gh/StrongerXi/73/orig 2025-12-04T08:53:58.7384873Z * [new branch] gh/XilunWu/160/base -> origin/gh/XilunWu/160/base 2025-12-04T08:53:58.7385144Z * [new branch] gh/XilunWu/160/head -> origin/gh/XilunWu/160/head 2025-12-04T08:53:58.7385342Z * [new branch] gh/XilunWu/160/orig -> origin/gh/XilunWu/160/orig 2025-12-04T08:53:58.7385532Z * [new branch] gh/XilunWu/163/base -> origin/gh/XilunWu/163/base 2025-12-04T08:53:58.7385725Z * [new branch] gh/XilunWu/163/head -> origin/gh/XilunWu/163/head 2025-12-04T08:53:58.7385906Z * [new branch] gh/XilunWu/163/orig -> origin/gh/XilunWu/163/orig 2025-12-04T08:53:58.7386094Z * [new branch] gh/XilunWu/168/base -> origin/gh/XilunWu/168/base 2025-12-04T08:53:58.7386283Z * [new branch] gh/XilunWu/168/head -> origin/gh/XilunWu/168/head 2025-12-04T08:53:58.7386463Z * [new branch] gh/XilunWu/168/orig -> origin/gh/XilunWu/168/orig 2025-12-04T08:53:58.7386652Z * [new branch] gh/XilunWu/169/base -> origin/gh/XilunWu/169/base 2025-12-04T08:53:58.7386880Z * [new branch] gh/XilunWu/169/head -> origin/gh/XilunWu/169/head 2025-12-04T08:53:58.7387065Z * [new branch] gh/XilunWu/169/orig -> origin/gh/XilunWu/169/orig 2025-12-04T08:53:58.7387246Z * [new branch] gh/XilunWu/170/base -> origin/gh/XilunWu/170/base 2025-12-04T08:53:58.7387471Z * [new branch] gh/XilunWu/170/head -> origin/gh/XilunWu/170/head 2025-12-04T08:53:58.7387667Z * [new branch] gh/XilunWu/170/orig -> origin/gh/XilunWu/170/orig 2025-12-04T08:53:58.7387855Z * [new branch] gh/XilunWu/171/base -> origin/gh/XilunWu/171/base 2025-12-04T08:53:58.7388039Z * [new branch] gh/XilunWu/171/head -> origin/gh/XilunWu/171/head 2025-12-04T08:53:58.7388216Z * [new branch] gh/XilunWu/171/orig -> origin/gh/XilunWu/171/orig 2025-12-04T08:53:58.7388396Z * [new branch] gh/XilunWu/173/base -> origin/gh/XilunWu/173/base 2025-12-04T08:53:58.7388579Z * [new branch] gh/XilunWu/173/head -> origin/gh/XilunWu/173/head 2025-12-04T08:53:58.7388763Z * [new branch] gh/XilunWu/173/orig -> origin/gh/XilunWu/173/orig 2025-12-04T08:53:58.7388949Z * [new branch] gh/XilunWu/175/base -> origin/gh/XilunWu/175/base 2025-12-04T08:53:58.7389133Z * [new branch] gh/XilunWu/175/head -> origin/gh/XilunWu/175/head 2025-12-04T08:53:58.7389315Z * [new branch] gh/XilunWu/175/orig -> origin/gh/XilunWu/175/orig 2025-12-04T08:53:58.7389500Z * [new branch] gh/XilunWu/176/base -> origin/gh/XilunWu/176/base 2025-12-04T08:53:58.7389684Z * [new branch] gh/XilunWu/176/head -> origin/gh/XilunWu/176/head 2025-12-04T08:53:58.7389871Z * [new branch] gh/XilunWu/176/orig -> origin/gh/XilunWu/176/orig 2025-12-04T08:53:58.7390060Z * [new branch] gh/XuehaiPan/14/base -> origin/gh/XuehaiPan/14/base 2025-12-04T08:53:58.7390251Z * [new branch] gh/XuehaiPan/14/head -> origin/gh/XuehaiPan/14/head 2025-12-04T08:53:58.7390436Z * [new branch] gh/XuehaiPan/14/orig -> origin/gh/XuehaiPan/14/orig 2025-12-04T08:53:58.7390631Z * [new branch] gh/XuehaiPan/179/base -> origin/gh/XuehaiPan/179/base 2025-12-04T08:53:58.7390828Z * [new branch] gh/XuehaiPan/179/head -> origin/gh/XuehaiPan/179/head 2025-12-04T08:53:58.7391024Z * [new branch] gh/XuehaiPan/179/orig -> origin/gh/XuehaiPan/179/orig 2025-12-04T08:53:58.7391218Z * [new branch] gh/XuehaiPan/249/base -> origin/gh/XuehaiPan/249/base 2025-12-04T08:53:58.7391408Z * [new branch] gh/XuehaiPan/249/head -> origin/gh/XuehaiPan/249/head 2025-12-04T08:53:58.7391606Z * [new branch] gh/XuehaiPan/249/orig -> origin/gh/XuehaiPan/249/orig 2025-12-04T08:53:58.7391845Z * [new branch] gh/XuehaiPan/253/base -> origin/gh/XuehaiPan/253/base 2025-12-04T08:53:58.7392035Z * [new branch] gh/XuehaiPan/253/head -> origin/gh/XuehaiPan/253/head 2025-12-04T08:53:58.7392228Z * [new branch] gh/XuehaiPan/253/orig -> origin/gh/XuehaiPan/253/orig 2025-12-04T08:53:58.7392417Z * [new branch] gh/XuehaiPan/254/base -> origin/gh/XuehaiPan/254/base 2025-12-04T08:53:58.7392610Z * [new branch] gh/XuehaiPan/254/head -> origin/gh/XuehaiPan/254/head 2025-12-04T08:53:58.7392804Z * [new branch] gh/XuehaiPan/254/orig -> origin/gh/XuehaiPan/254/orig 2025-12-04T08:53:58.7392993Z * [new branch] gh/XuehaiPan/255/base -> origin/gh/XuehaiPan/255/base 2025-12-04T08:53:58.7393186Z * [new branch] gh/XuehaiPan/255/head -> origin/gh/XuehaiPan/255/head 2025-12-04T08:53:58.7393380Z * [new branch] gh/XuehaiPan/255/orig -> origin/gh/XuehaiPan/255/orig 2025-12-04T08:53:58.7393571Z * [new branch] gh/XuehaiPan/271/base -> origin/gh/XuehaiPan/271/base 2025-12-04T08:53:58.7393764Z * [new branch] gh/XuehaiPan/271/head -> origin/gh/XuehaiPan/271/head 2025-12-04T08:53:58.7393966Z * [new branch] gh/XuehaiPan/271/orig -> origin/gh/XuehaiPan/271/orig 2025-12-04T08:53:58.7394159Z * [new branch] gh/XuehaiPan/343/base -> origin/gh/XuehaiPan/343/base 2025-12-04T08:53:58.7394383Z * [new branch] gh/XuehaiPan/343/head -> origin/gh/XuehaiPan/343/head 2025-12-04T08:53:58.7394580Z * [new branch] gh/XuehaiPan/343/orig -> origin/gh/XuehaiPan/343/orig 2025-12-04T08:53:58.7394768Z * [new branch] gh/XuehaiPan/347/base -> origin/gh/XuehaiPan/347/base 2025-12-04T08:53:58.7394962Z * [new branch] gh/XuehaiPan/347/head -> origin/gh/XuehaiPan/347/head 2025-12-04T08:53:58.7395155Z * [new branch] gh/XuehaiPan/347/orig -> origin/gh/XuehaiPan/347/orig 2025-12-04T08:53:58.7395348Z * [new branch] gh/XuehaiPan/348/base -> origin/gh/XuehaiPan/348/base 2025-12-04T08:53:58.7395541Z * [new branch] gh/XuehaiPan/348/head -> origin/gh/XuehaiPan/348/head 2025-12-04T08:53:58.7395735Z * [new branch] gh/XuehaiPan/348/orig -> origin/gh/XuehaiPan/348/orig 2025-12-04T08:53:58.7395927Z * [new branch] gh/XuehaiPan/350/base -> origin/gh/XuehaiPan/350/base 2025-12-04T08:53:58.7396123Z * [new branch] gh/XuehaiPan/350/head -> origin/gh/XuehaiPan/350/head 2025-12-04T08:53:58.7396318Z * [new branch] gh/XuehaiPan/350/orig -> origin/gh/XuehaiPan/350/orig 2025-12-04T08:53:58.7396508Z * [new branch] gh/XuehaiPan/365/base -> origin/gh/XuehaiPan/365/base 2025-12-04T08:53:58.7396703Z * [new branch] gh/XuehaiPan/365/head -> origin/gh/XuehaiPan/365/head 2025-12-04T08:53:58.7396952Z * [new branch] gh/XuehaiPan/365/orig -> origin/gh/XuehaiPan/365/orig 2025-12-04T08:53:58.7397145Z * [new branch] gh/XuehaiPan/366/base -> origin/gh/XuehaiPan/366/base 2025-12-04T08:53:58.7397343Z * [new branch] gh/XuehaiPan/366/head -> origin/gh/XuehaiPan/366/head 2025-12-04T08:53:58.7397535Z * [new branch] gh/XuehaiPan/370/base -> origin/gh/XuehaiPan/370/base 2025-12-04T08:53:58.7397734Z * [new branch] gh/XuehaiPan/370/head -> origin/gh/XuehaiPan/370/head 2025-12-04T08:53:58.7397929Z * [new branch] gh/XuehaiPan/370/orig -> origin/gh/XuehaiPan/370/orig 2025-12-04T08:53:58.7398118Z * [new branch] gh/XuehaiPan/390/base -> origin/gh/XuehaiPan/390/base 2025-12-04T08:53:58.7398314Z * [new branch] gh/XuehaiPan/390/head -> origin/gh/XuehaiPan/390/head 2025-12-04T08:53:58.7398510Z * [new branch] gh/XuehaiPan/390/orig -> origin/gh/XuehaiPan/390/orig 2025-12-04T08:53:58.7398699Z * [new branch] gh/XuehaiPan/391/base -> origin/gh/XuehaiPan/391/base 2025-12-04T08:53:58.7398934Z * [new branch] gh/XuehaiPan/391/head -> origin/gh/XuehaiPan/391/head 2025-12-04T08:53:58.7399129Z * [new branch] gh/XuehaiPan/391/orig -> origin/gh/XuehaiPan/391/orig 2025-12-04T08:53:58.7399319Z * [new branch] gh/XuehaiPan/392/base -> origin/gh/XuehaiPan/392/base 2025-12-04T08:53:58.7399514Z * [new branch] gh/XuehaiPan/392/head -> origin/gh/XuehaiPan/392/head 2025-12-04T08:53:58.7399708Z * [new branch] gh/XuehaiPan/392/orig -> origin/gh/XuehaiPan/392/orig 2025-12-04T08:53:58.7399899Z * [new branch] gh/XuehaiPan/394/base -> origin/gh/XuehaiPan/394/base 2025-12-04T08:53:58.7400093Z * [new branch] gh/XuehaiPan/394/head -> origin/gh/XuehaiPan/394/head 2025-12-04T08:53:58.7400287Z * [new branch] gh/XuehaiPan/394/orig -> origin/gh/XuehaiPan/394/orig 2025-12-04T08:53:58.7400478Z * [new branch] gh/XuehaiPan/397/base -> origin/gh/XuehaiPan/397/base 2025-12-04T08:53:58.7400674Z * [new branch] gh/XuehaiPan/397/head -> origin/gh/XuehaiPan/397/head 2025-12-04T08:53:58.7400867Z * [new branch] gh/XuehaiPan/397/orig -> origin/gh/XuehaiPan/397/orig 2025-12-04T08:53:58.7401058Z * [new branch] gh/XuehaiPan/398/base -> origin/gh/XuehaiPan/398/base 2025-12-04T08:53:58.7401288Z * [new branch] gh/XuehaiPan/398/head -> origin/gh/XuehaiPan/398/head 2025-12-04T08:53:58.7401483Z * [new branch] gh/XuehaiPan/398/orig -> origin/gh/XuehaiPan/398/orig 2025-12-04T08:53:58.7401674Z * [new branch] gh/XuehaiPan/399/base -> origin/gh/XuehaiPan/399/base 2025-12-04T08:53:58.7401869Z * [new branch] gh/XuehaiPan/399/head -> origin/gh/XuehaiPan/399/head 2025-12-04T08:53:58.7402058Z * [new branch] gh/XuehaiPan/399/orig -> origin/gh/XuehaiPan/399/orig 2025-12-04T08:53:58.7402257Z * [new branch] gh/XuehaiPan/400/base -> origin/gh/XuehaiPan/400/base 2025-12-04T08:53:58.7402452Z * [new branch] gh/XuehaiPan/400/head -> origin/gh/XuehaiPan/400/head 2025-12-04T08:53:58.7402642Z * [new branch] gh/XuehaiPan/400/orig -> origin/gh/XuehaiPan/400/orig 2025-12-04T08:53:58.7402843Z * [new branch] gh/ZhiweiYan-96/39/base -> origin/gh/ZhiweiYan-96/39/base 2025-12-04T08:53:58.7403047Z * [new branch] gh/ZhiweiYan-96/39/head -> origin/gh/ZhiweiYan-96/39/head 2025-12-04T08:53:58.7403243Z * [new branch] gh/ZhiweiYan-96/39/orig -> origin/gh/ZhiweiYan-96/39/orig 2025-12-04T08:53:58.7403441Z * [new branch] gh/ZhiweiYan-96/44/base -> origin/gh/ZhiweiYan-96/44/base 2025-12-04T08:53:58.7403639Z * [new branch] gh/ZhiweiYan-96/44/head -> origin/gh/ZhiweiYan-96/44/head 2025-12-04T08:53:58.7403833Z * [new branch] gh/ZhiweiYan-96/45/base -> origin/gh/ZhiweiYan-96/45/base 2025-12-04T08:53:58.7404032Z * [new branch] gh/ZhiweiYan-96/45/head -> origin/gh/ZhiweiYan-96/45/head 2025-12-04T08:53:58.7404228Z * [new branch] gh/ZhiweiYan-96/49/base -> origin/gh/ZhiweiYan-96/49/base 2025-12-04T08:53:58.7404421Z * [new branch] gh/ZhiweiYan-96/49/head -> origin/gh/ZhiweiYan-96/49/head 2025-12-04T08:53:58.7404618Z * [new branch] gh/ZhiweiYan-96/62/base -> origin/gh/ZhiweiYan-96/62/base 2025-12-04T08:53:58.7404813Z * [new branch] gh/ZhiweiYan-96/62/head -> origin/gh/ZhiweiYan-96/62/head 2025-12-04T08:53:58.7405005Z * [new branch] gh/ZhiweiYan-96/66/base -> origin/gh/ZhiweiYan-96/66/base 2025-12-04T08:53:58.7405202Z * [new branch] gh/ZhiweiYan-96/66/head -> origin/gh/ZhiweiYan-96/66/head 2025-12-04T08:53:58.7405398Z * [new branch] gh/ZhiweiYan-96/67/base -> origin/gh/ZhiweiYan-96/67/base 2025-12-04T08:53:58.7405591Z * [new branch] gh/ZhiweiYan-96/67/head -> origin/gh/ZhiweiYan-96/67/head 2025-12-04T08:53:58.7405813Z * [new branch] gh/ZhiweiYan-96/68/base -> origin/gh/ZhiweiYan-96/68/base 2025-12-04T08:53:58.7406009Z * [new branch] gh/ZhiweiYan-96/68/head -> origin/gh/ZhiweiYan-96/68/head 2025-12-04T08:53:58.7406200Z * [new branch] gh/ZhiweiYan-96/68/orig -> origin/gh/ZhiweiYan-96/68/orig 2025-12-04T08:53:58.7406399Z * [new branch] gh/aakhundov/1/base -> origin/gh/aakhundov/1/base 2025-12-04T08:53:58.7406587Z * [new branch] gh/aakhundov/1/head -> origin/gh/aakhundov/1/head 2025-12-04T08:53:58.7406831Z * [new branch] gh/aakhundov/2/base -> origin/gh/aakhundov/2/base 2025-12-04T08:53:58.7407026Z * [new branch] gh/aakhundov/2/head -> origin/gh/aakhundov/2/head 2025-12-04T08:53:58.7407218Z * [new branch] gh/aditew01/openblas -> origin/gh/aditew01/openblas 2025-12-04T08:53:58.7407418Z * [new branch] gh/aditew01/sbgemm -> origin/gh/aditew01/sbgemm 2025-12-04T08:53:58.7407613Z * [new branch] gh/aditew01/vecbf16 -> origin/gh/aditew01/vecbf16 2025-12-04T08:53:58.7407799Z * [new branch] gh/albanD/4/base -> origin/gh/albanD/4/base 2025-12-04T08:53:58.7407983Z * [new branch] gh/albanD/4/head -> origin/gh/albanD/4/head 2025-12-04T08:53:58.7408208Z * [new branch] gh/albanD/4/orig -> origin/gh/albanD/4/orig 2025-12-04T08:53:58.7408481Z * [new branch] gh/alexbrauckmann/paddedtensor_faketensor_init -> origin/gh/alexbrauckmann/paddedtensor_faketensor_init 2025-12-04T08:53:58.7408760Z * [new branch] gh/alexsamardzic/12/base -> origin/gh/alexsamardzic/12/base 2025-12-04T08:53:58.7408971Z * [new branch] gh/alexsamardzic/12/head -> origin/gh/alexsamardzic/12/head 2025-12-04T08:53:58.7409176Z * [new branch] gh/alexsamardzic/12/orig -> origin/gh/alexsamardzic/12/orig 2025-12-04T08:53:58.7409388Z * [new branch] gh/alexsamardzic/14/base -> origin/gh/alexsamardzic/14/base 2025-12-04T08:53:58.7409595Z * [new branch] gh/alexsamardzic/14/head -> origin/gh/alexsamardzic/14/head 2025-12-04T08:53:58.7409796Z * [new branch] gh/alexsamardzic/14/orig -> origin/gh/alexsamardzic/14/orig 2025-12-04T08:53:58.7410007Z * [new branch] gh/alexsamardzic/15/base -> origin/gh/alexsamardzic/15/base 2025-12-04T08:53:58.7410215Z * [new branch] gh/alexsamardzic/15/head -> origin/gh/alexsamardzic/15/head 2025-12-04T08:53:58.7410417Z * [new branch] gh/alexsamardzic/15/orig -> origin/gh/alexsamardzic/15/orig 2025-12-04T08:53:58.7410616Z * [new branch] gh/amjames/18/base -> origin/gh/amjames/18/base 2025-12-04T08:53:58.7410806Z * [new branch] gh/amjames/18/head -> origin/gh/amjames/18/head 2025-12-04T08:53:58.7410990Z * [new branch] gh/amjames/18/orig -> origin/gh/amjames/18/orig 2025-12-04T08:53:58.7411185Z * [new branch] gh/andrewor14/35/base -> origin/gh/andrewor14/35/base 2025-12-04T08:53:58.7411381Z * [new branch] gh/andrewor14/35/head -> origin/gh/andrewor14/35/head 2025-12-04T08:53:58.7411571Z * [new branch] gh/andrewor14/35/orig -> origin/gh/andrewor14/35/orig 2025-12-04T08:53:58.7411763Z * [new branch] gh/andrewor14/50/base -> origin/gh/andrewor14/50/base 2025-12-04T08:53:58.7411949Z * [new branch] gh/andrewor14/50/head -> origin/gh/andrewor14/50/head 2025-12-04T08:53:58.7412140Z * [new branch] gh/andrewor14/50/orig -> origin/gh/andrewor14/50/orig 2025-12-04T08:53:58.7412330Z * [new branch] gh/andyanwang/30/base -> origin/gh/andyanwang/30/base 2025-12-04T08:53:58.7412519Z * [new branch] gh/andyanwang/30/orig -> origin/gh/andyanwang/30/orig 2025-12-04T08:53:58.7412712Z * [new branch] gh/andyanwang/31/base -> origin/gh/andyanwang/31/base 2025-12-04T08:53:58.7412939Z * [new branch] gh/andyanwang/31/orig -> origin/gh/andyanwang/31/orig 2025-12-04T08:53:58.7413128Z * [new branch] gh/andyanwang/39/base -> origin/gh/andyanwang/39/base 2025-12-04T08:53:58.7413322Z * [new branch] gh/andyanwang/39/head -> origin/gh/andyanwang/39/head 2025-12-04T08:53:58.7413518Z * [new branch] gh/andyanwang/39/orig -> origin/gh/andyanwang/39/orig 2025-12-04T08:53:58.7413708Z * [new branch] gh/andyanwang/42/base -> origin/gh/andyanwang/42/base 2025-12-04T08:53:58.7413902Z * [new branch] gh/andyanwang/42/head -> origin/gh/andyanwang/42/head 2025-12-04T08:53:58.7414095Z * [new branch] gh/andyanwang/42/orig -> origin/gh/andyanwang/42/orig 2025-12-04T08:53:58.7414283Z * [new branch] gh/andyanwang/45/base -> origin/gh/andyanwang/45/base 2025-12-04T08:53:58.7414481Z * [new branch] gh/andyanwang/45/head -> origin/gh/andyanwang/45/head 2025-12-04T08:53:58.7414674Z * [new branch] gh/andyanwang/45/orig -> origin/gh/andyanwang/45/orig 2025-12-04T08:53:58.7414865Z * [new branch] gh/angelayi/107/base -> origin/gh/angelayi/107/base 2025-12-04T08:53:58.7415056Z * [new branch] gh/angelayi/107/head -> origin/gh/angelayi/107/head 2025-12-04T08:53:58.7415276Z * [new branch] gh/angelayi/114/base -> origin/gh/angelayi/114/base 2025-12-04T08:53:58.7415465Z * [new branch] gh/angelayi/114/head -> origin/gh/angelayi/114/head 2025-12-04T08:53:58.7415655Z * [new branch] gh/angelayi/114/orig -> origin/gh/angelayi/114/orig 2025-12-04T08:53:58.7415846Z * [new branch] gh/angelayi/116/base -> origin/gh/angelayi/116/base 2025-12-04T08:53:58.7416030Z * [new branch] gh/angelayi/116/head -> origin/gh/angelayi/116/head 2025-12-04T08:53:58.7416224Z * [new branch] gh/angelayi/116/orig -> origin/gh/angelayi/116/orig 2025-12-04T08:53:58.7416408Z * [new branch] gh/angelayi/122/base -> origin/gh/angelayi/122/base 2025-12-04T08:53:58.7416596Z * [new branch] gh/angelayi/122/head -> origin/gh/angelayi/122/head 2025-12-04T08:53:58.7416837Z * [new branch] gh/angelayi/122/orig -> origin/gh/angelayi/122/orig 2025-12-04T08:53:58.7417024Z * [new branch] gh/angelayi/124/base -> origin/gh/angelayi/124/base 2025-12-04T08:53:58.7417210Z * [new branch] gh/angelayi/124/head -> origin/gh/angelayi/124/head 2025-12-04T08:53:58.7417393Z * [new branch] gh/angelayi/124/orig -> origin/gh/angelayi/124/orig 2025-12-04T08:53:58.7417576Z * [new branch] gh/angelayi/128/base -> origin/gh/angelayi/128/base 2025-12-04T08:53:58.7417764Z * [new branch] gh/angelayi/128/head -> origin/gh/angelayi/128/head 2025-12-04T08:53:58.7417953Z * [new branch] gh/angelayi/128/orig -> origin/gh/angelayi/128/orig 2025-12-04T08:53:58.7418140Z * [new branch] gh/angelayi/131/base -> origin/gh/angelayi/131/base 2025-12-04T08:53:58.7418328Z * [new branch] gh/angelayi/131/head -> origin/gh/angelayi/131/head 2025-12-04T08:53:58.7418513Z * [new branch] gh/angelayi/131/orig -> origin/gh/angelayi/131/orig 2025-12-04T08:53:58.7418699Z * [new branch] gh/angelayi/132/base -> origin/gh/angelayi/132/base 2025-12-04T08:53:58.7418884Z * [new branch] gh/angelayi/132/head -> origin/gh/angelayi/132/head 2025-12-04T08:53:58.7419071Z * [new branch] gh/angelayi/132/orig -> origin/gh/angelayi/132/orig 2025-12-04T08:53:58.7419253Z * [new branch] gh/angelayi/133/base -> origin/gh/angelayi/133/base 2025-12-04T08:53:58.7419440Z * [new branch] gh/angelayi/133/head -> origin/gh/angelayi/133/head 2025-12-04T08:53:58.7419669Z * [new branch] gh/angelayi/133/orig -> origin/gh/angelayi/133/orig 2025-12-04T08:53:58.7419852Z * [new branch] gh/angelayi/134/base -> origin/gh/angelayi/134/base 2025-12-04T08:53:58.7420036Z * [new branch] gh/angelayi/134/head -> origin/gh/angelayi/134/head 2025-12-04T08:53:58.7420221Z * [new branch] gh/angelayi/134/orig -> origin/gh/angelayi/134/orig 2025-12-04T08:53:58.7420406Z * [new branch] gh/angelayi/135/base -> origin/gh/angelayi/135/base 2025-12-04T08:53:58.7420591Z * [new branch] gh/angelayi/135/head -> origin/gh/angelayi/135/head 2025-12-04T08:53:58.7420774Z * [new branch] gh/angelayi/135/orig -> origin/gh/angelayi/135/orig 2025-12-04T08:53:58.7420958Z * [new branch] gh/angelayi/136/base -> origin/gh/angelayi/136/base 2025-12-04T08:53:58.7421142Z * [new branch] gh/angelayi/136/head -> origin/gh/angelayi/136/head 2025-12-04T08:53:58.7421326Z * [new branch] gh/angelayi/136/orig -> origin/gh/angelayi/136/orig 2025-12-04T08:53:58.7421511Z * [new branch] gh/angelayi/137/base -> origin/gh/angelayi/137/base 2025-12-04T08:53:58.7421699Z * [new branch] gh/angelayi/137/head -> origin/gh/angelayi/137/head 2025-12-04T08:53:58.7421910Z * [new branch] gh/angelayi/137/orig -> origin/gh/angelayi/137/orig 2025-12-04T08:53:58.7422098Z * [new branch] gh/angelayi/138/base -> origin/gh/angelayi/138/base 2025-12-04T08:53:58.7422282Z * [new branch] gh/angelayi/138/head -> origin/gh/angelayi/138/head 2025-12-04T08:53:58.7422464Z * [new branch] gh/angelayi/138/orig -> origin/gh/angelayi/138/orig 2025-12-04T08:53:58.7422650Z * [new branch] gh/angelayi/139/base -> origin/gh/angelayi/139/base 2025-12-04T08:53:58.7422836Z * [new branch] gh/angelayi/139/head -> origin/gh/angelayi/139/head 2025-12-04T08:53:58.7423022Z * [new branch] gh/angelayi/139/orig -> origin/gh/angelayi/139/orig 2025-12-04T08:53:58.7423207Z * [new branch] gh/angelayi/140/base -> origin/gh/angelayi/140/base 2025-12-04T08:53:58.7423394Z * [new branch] gh/angelayi/140/head -> origin/gh/angelayi/140/head 2025-12-04T08:53:58.7423581Z * [new branch] gh/angelayi/140/orig -> origin/gh/angelayi/140/orig 2025-12-04T08:53:58.7423763Z * [new branch] gh/angelayi/141/base -> origin/gh/angelayi/141/base 2025-12-04T08:53:58.7423948Z * [new branch] gh/angelayi/141/head -> origin/gh/angelayi/141/head 2025-12-04T08:53:58.7424128Z * [new branch] gh/angelayi/141/orig -> origin/gh/angelayi/141/orig 2025-12-04T08:53:58.7424309Z * [new branch] gh/angelayi/142/base -> origin/gh/angelayi/142/base 2025-12-04T08:53:58.7424494Z * [new branch] gh/angelayi/142/head -> origin/gh/angelayi/142/head 2025-12-04T08:53:58.7424675Z * [new branch] gh/angelayi/142/orig -> origin/gh/angelayi/142/orig 2025-12-04T08:53:58.7424857Z * [new branch] gh/angelayi/143/base -> origin/gh/angelayi/143/base 2025-12-04T08:53:58.7425040Z * [new branch] gh/angelayi/143/head -> origin/gh/angelayi/143/head 2025-12-04T08:53:58.7425221Z * [new branch] gh/angelayi/143/orig -> origin/gh/angelayi/143/orig 2025-12-04T08:53:58.7425461Z * [new branch] gh/angelayi/144/base -> origin/gh/angelayi/144/base 2025-12-04T08:53:58.7425648Z * [new branch] gh/angelayi/144/head -> origin/gh/angelayi/144/head 2025-12-04T08:53:58.7425828Z * [new branch] gh/angelayi/144/orig -> origin/gh/angelayi/144/orig 2025-12-04T08:53:58.7426016Z * [new branch] gh/anijain2305/753/base -> origin/gh/anijain2305/753/base 2025-12-04T08:53:58.7426209Z * [new branch] gh/anijain2305/753/head -> origin/gh/anijain2305/753/head 2025-12-04T08:53:58.7426429Z * [new branch] gh/anijain2305/753/orig -> origin/gh/anijain2305/753/orig 2025-12-04T08:53:58.7426618Z * [new branch] gh/anijain2305/810/base -> origin/gh/anijain2305/810/base 2025-12-04T08:53:58.7426852Z * [new branch] gh/anijain2305/810/head -> origin/gh/anijain2305/810/head 2025-12-04T08:53:58.7427044Z * [new branch] gh/anijain2305/810/orig -> origin/gh/anijain2305/810/orig 2025-12-04T08:53:58.7427238Z * [new branch] gh/anijain2305/854/base -> origin/gh/anijain2305/854/base 2025-12-04T08:53:58.7427429Z * [new branch] gh/anijain2305/854/head -> origin/gh/anijain2305/854/head 2025-12-04T08:53:58.7427621Z * [new branch] gh/anijain2305/854/orig -> origin/gh/anijain2305/854/orig 2025-12-04T08:53:58.7427809Z * [new branch] gh/anijain2305/864/base -> origin/gh/anijain2305/864/base 2025-12-04T08:53:58.7428001Z * [new branch] gh/anijain2305/864/head -> origin/gh/anijain2305/864/head 2025-12-04T08:53:58.7428189Z * [new branch] gh/anijain2305/864/orig -> origin/gh/anijain2305/864/orig 2025-12-04T08:53:58.7428376Z * [new branch] gh/anijain2305/870/base -> origin/gh/anijain2305/870/base 2025-12-04T08:53:58.7428564Z * [new branch] gh/anijain2305/870/head -> origin/gh/anijain2305/870/head 2025-12-04T08:53:58.7428793Z * [new branch] gh/anijain2305/870/orig -> origin/gh/anijain2305/870/orig 2025-12-04T08:53:58.7428982Z * [new branch] gh/anijain2305/873/base -> origin/gh/anijain2305/873/base 2025-12-04T08:53:58.7429169Z * [new branch] gh/anijain2305/873/head -> origin/gh/anijain2305/873/head 2025-12-04T08:53:58.7429355Z * [new branch] gh/anijain2305/873/orig -> origin/gh/anijain2305/873/orig 2025-12-04T08:53:58.7429545Z * [new branch] gh/anijain2305/894/base -> origin/gh/anijain2305/894/base 2025-12-04T08:53:58.7429735Z * [new branch] gh/anijain2305/894/head -> origin/gh/anijain2305/894/head 2025-12-04T08:53:58.7429923Z * [new branch] gh/anijain2305/894/orig -> origin/gh/anijain2305/894/orig 2025-12-04T08:53:58.7430113Z * [new branch] gh/anijain2305/895/base -> origin/gh/anijain2305/895/base 2025-12-04T08:53:58.7430302Z * [new branch] gh/anijain2305/895/head -> origin/gh/anijain2305/895/head 2025-12-04T08:53:58.7430493Z * [new branch] gh/anijain2305/895/orig -> origin/gh/anijain2305/895/orig 2025-12-04T08:53:58.7430685Z * [new branch] gh/anijain2305/910/base -> origin/gh/anijain2305/910/base 2025-12-04T08:53:58.7430872Z * [new branch] gh/anijain2305/910/head -> origin/gh/anijain2305/910/head 2025-12-04T08:53:58.7431065Z * [new branch] gh/anijain2305/910/orig -> origin/gh/anijain2305/910/orig 2025-12-04T08:53:58.7431252Z * [new branch] gh/anijain2305/919/base -> origin/gh/anijain2305/919/base 2025-12-04T08:53:58.7431443Z * [new branch] gh/anijain2305/919/head -> origin/gh/anijain2305/919/head 2025-12-04T08:53:58.7431632Z * [new branch] gh/anijain2305/919/orig -> origin/gh/anijain2305/919/orig 2025-12-04T08:53:58.7431821Z * [new branch] gh/anijain2305/922/base -> origin/gh/anijain2305/922/base 2025-12-04T08:53:58.7432019Z * [new branch] gh/anijain2305/922/head -> origin/gh/anijain2305/922/head 2025-12-04T08:53:58.7432212Z * [new branch] gh/anijain2305/922/orig -> origin/gh/anijain2305/922/orig 2025-12-04T08:53:58.7432400Z * [new branch] gh/anijain2305/932/base -> origin/gh/anijain2305/932/base 2025-12-04T08:53:58.7432592Z * [new branch] gh/anijain2305/932/head -> origin/gh/anijain2305/932/head 2025-12-04T08:53:58.7432782Z * [new branch] gh/anijain2305/932/orig -> origin/gh/anijain2305/932/orig 2025-12-04T08:53:58.7433005Z * [new branch] gh/anijain2305/940/base -> origin/gh/anijain2305/940/base 2025-12-04T08:53:58.7433195Z * [new branch] gh/anijain2305/940/head -> origin/gh/anijain2305/940/head 2025-12-04T08:53:58.7433388Z * [new branch] gh/anijain2305/940/orig -> origin/gh/anijain2305/940/orig 2025-12-04T08:53:58.7433576Z * [new branch] gh/anijain2305/941/base -> origin/gh/anijain2305/941/base 2025-12-04T08:53:58.7433769Z * [new branch] gh/anijain2305/941/head -> origin/gh/anijain2305/941/head 2025-12-04T08:53:58.7433962Z * [new branch] gh/anijain2305/941/orig -> origin/gh/anijain2305/941/orig 2025-12-04T08:53:58.7434151Z * [new branch] gh/anijain2305/942/base -> origin/gh/anijain2305/942/base 2025-12-04T08:53:58.7434345Z * [new branch] gh/anijain2305/942/head -> origin/gh/anijain2305/942/head 2025-12-04T08:53:58.7434533Z * [new branch] gh/anijain2305/942/orig -> origin/gh/anijain2305/942/orig 2025-12-04T08:53:58.7434723Z * [new branch] gh/anijain2305/943/base -> origin/gh/anijain2305/943/base 2025-12-04T08:53:58.7434915Z * [new branch] gh/anijain2305/943/head -> origin/gh/anijain2305/943/head 2025-12-04T08:53:58.7435108Z * [new branch] gh/anijain2305/943/orig -> origin/gh/anijain2305/943/orig 2025-12-04T08:53:58.7435326Z * [new branch] gh/anijain2305/944/base -> origin/gh/anijain2305/944/base 2025-12-04T08:53:58.7435518Z * [new branch] gh/anijain2305/944/head -> origin/gh/anijain2305/944/head 2025-12-04T08:53:58.7435707Z * [new branch] gh/anijain2305/944/orig -> origin/gh/anijain2305/944/orig 2025-12-04T08:53:58.7435901Z * [new branch] gh/anijain2305/945/base -> origin/gh/anijain2305/945/base 2025-12-04T08:53:58.7436093Z * [new branch] gh/anijain2305/945/head -> origin/gh/anijain2305/945/head 2025-12-04T08:53:58.7436281Z * [new branch] gh/anijain2305/945/orig -> origin/gh/anijain2305/945/orig 2025-12-04T08:53:58.7436478Z * [new branch] gh/anijain2305/946/base -> origin/gh/anijain2305/946/base 2025-12-04T08:53:58.7436674Z * [new branch] gh/anijain2305/946/head -> origin/gh/anijain2305/946/head 2025-12-04T08:53:58.7436910Z * [new branch] gh/anijain2305/946/orig -> origin/gh/anijain2305/946/orig 2025-12-04T08:53:58.7437108Z * [new branch] gh/anijain2305/947/base -> origin/gh/anijain2305/947/base 2025-12-04T08:53:58.7437301Z * [new branch] gh/anijain2305/947/head -> origin/gh/anijain2305/947/head 2025-12-04T08:53:58.7437493Z * [new branch] gh/anijain2305/947/orig -> origin/gh/anijain2305/947/orig 2025-12-04T08:53:58.7437684Z * [new branch] gh/anijain2305/948/base -> origin/gh/anijain2305/948/base 2025-12-04T08:53:58.7437877Z * [new branch] gh/anijain2305/948/head -> origin/gh/anijain2305/948/head 2025-12-04T08:53:58.7438069Z * [new branch] gh/anijain2305/948/orig -> origin/gh/anijain2305/948/orig 2025-12-04T08:53:58.7438261Z * [new branch] gh/anijain2305/949/base -> origin/gh/anijain2305/949/base 2025-12-04T08:53:58.7438452Z * [new branch] gh/anijain2305/949/head -> origin/gh/anijain2305/949/head 2025-12-04T08:53:58.7438643Z * [new branch] gh/anijain2305/949/orig -> origin/gh/anijain2305/949/orig 2025-12-04T08:53:58.7438838Z * [new branch] gh/anijain2305/950/base -> origin/gh/anijain2305/950/base 2025-12-04T08:53:58.7439031Z * [new branch] gh/anijain2305/950/head -> origin/gh/anijain2305/950/head 2025-12-04T08:53:58.7439221Z * [new branch] gh/anijain2305/950/orig -> origin/gh/anijain2305/950/orig 2025-12-04T08:53:58.7439414Z * [new branch] gh/anijain2305/951/base -> origin/gh/anijain2305/951/base 2025-12-04T08:53:58.7439607Z * [new branch] gh/anijain2305/951/head -> origin/gh/anijain2305/951/head 2025-12-04T08:53:58.7439844Z * [new branch] gh/anijain2305/951/orig -> origin/gh/anijain2305/951/orig 2025-12-04T08:53:58.7440033Z * [new branch] gh/anijain2305/952/base -> origin/gh/anijain2305/952/base 2025-12-04T08:53:58.7440222Z * [new branch] gh/anijain2305/952/head -> origin/gh/anijain2305/952/head 2025-12-04T08:53:58.7440413Z * [new branch] gh/anijain2305/952/orig -> origin/gh/anijain2305/952/orig 2025-12-04T08:53:58.7440604Z * [new branch] gh/anijain2305/953/base -> origin/gh/anijain2305/953/base 2025-12-04T08:53:58.7440791Z * [new branch] gh/anijain2305/953/head -> origin/gh/anijain2305/953/head 2025-12-04T08:53:58.7440983Z * [new branch] gh/anijain2305/953/orig -> origin/gh/anijain2305/953/orig 2025-12-04T08:53:58.7441177Z * [new branch] gh/anijain2305/954/base -> origin/gh/anijain2305/954/base 2025-12-04T08:53:58.7441364Z * [new branch] gh/anijain2305/954/head -> origin/gh/anijain2305/954/head 2025-12-04T08:53:58.7441558Z * [new branch] gh/anijain2305/954/orig -> origin/gh/anijain2305/954/orig 2025-12-04T08:53:58.7441751Z * [new branch] gh/anijain2305/955/base -> origin/gh/anijain2305/955/base 2025-12-04T08:53:58.7441940Z * [new branch] gh/anijain2305/955/head -> origin/gh/anijain2305/955/head 2025-12-04T08:53:58.7442200Z * [new branch] gh/anijain2305/955/orig -> origin/gh/anijain2305/955/orig 2025-12-04T08:53:58.7442390Z * [new branch] gh/anijain2305/956/base -> origin/gh/anijain2305/956/base 2025-12-04T08:53:58.7442578Z * [new branch] gh/anijain2305/956/head -> origin/gh/anijain2305/956/head 2025-12-04T08:53:58.7442770Z * [new branch] gh/anijain2305/956/orig -> origin/gh/anijain2305/956/orig 2025-12-04T08:53:58.7442960Z * [new branch] gh/anijain2305/957/base -> origin/gh/anijain2305/957/base 2025-12-04T08:53:58.7443150Z * [new branch] gh/anijain2305/957/head -> origin/gh/anijain2305/957/head 2025-12-04T08:53:58.7443341Z * [new branch] gh/anijain2305/957/orig -> origin/gh/anijain2305/957/orig 2025-12-04T08:53:58.7443534Z * [new branch] gh/anijain2305/958/base -> origin/gh/anijain2305/958/base 2025-12-04T08:53:58.7443724Z * [new branch] gh/anijain2305/958/head -> origin/gh/anijain2305/958/head 2025-12-04T08:53:58.7443918Z * [new branch] gh/anijain2305/958/orig -> origin/gh/anijain2305/958/orig 2025-12-04T08:53:58.7444111Z * [new branch] gh/anijain2305/959/base -> origin/gh/anijain2305/959/base 2025-12-04T08:53:58.7444300Z * [new branch] gh/anijain2305/959/head -> origin/gh/anijain2305/959/head 2025-12-04T08:53:58.7444494Z * [new branch] gh/anijain2305/959/orig -> origin/gh/anijain2305/959/orig 2025-12-04T08:53:58.7444681Z * [new branch] gh/anijain2305/960/base -> origin/gh/anijain2305/960/base 2025-12-04T08:53:58.7444872Z * [new branch] gh/anijain2305/960/head -> origin/gh/anijain2305/960/head 2025-12-04T08:53:58.7445066Z * [new branch] gh/anijain2305/960/orig -> origin/gh/anijain2305/960/orig 2025-12-04T08:53:58.7445256Z * [new branch] gh/anijain2305/961/base -> origin/gh/anijain2305/961/base 2025-12-04T08:53:58.7445452Z * [new branch] gh/anijain2305/961/head -> origin/gh/anijain2305/961/head 2025-12-04T08:53:58.7445644Z * [new branch] gh/anijain2305/961/orig -> origin/gh/anijain2305/961/orig 2025-12-04T08:53:58.7445835Z * [new branch] gh/anijain2305/962/base -> origin/gh/anijain2305/962/base 2025-12-04T08:53:58.7446027Z * [new branch] gh/anijain2305/962/head -> origin/gh/anijain2305/962/head 2025-12-04T08:53:58.7446220Z * [new branch] gh/anijain2305/962/orig -> origin/gh/anijain2305/962/orig 2025-12-04T08:53:58.7446411Z * [new branch] gh/anijain2305/963/base -> origin/gh/anijain2305/963/base 2025-12-04T08:53:58.7446637Z * [new branch] gh/anijain2305/963/head -> origin/gh/anijain2305/963/head 2025-12-04T08:53:58.7446874Z * [new branch] gh/anijain2305/963/orig -> origin/gh/anijain2305/963/orig 2025-12-04T08:53:58.7447066Z * [new branch] gh/anijain2305/964/base -> origin/gh/anijain2305/964/base 2025-12-04T08:53:58.7447259Z * [new branch] gh/anijain2305/964/head -> origin/gh/anijain2305/964/head 2025-12-04T08:53:58.7447452Z * [new branch] gh/anijain2305/964/orig -> origin/gh/anijain2305/964/orig 2025-12-04T08:53:58.7447644Z * [new branch] gh/anijain2305/965/base -> origin/gh/anijain2305/965/base 2025-12-04T08:53:58.7447838Z * [new branch] gh/anijain2305/965/head -> origin/gh/anijain2305/965/head 2025-12-04T08:53:58.7448030Z * [new branch] gh/anijain2305/965/orig -> origin/gh/anijain2305/965/orig 2025-12-04T08:53:58.7448218Z * [new branch] gh/anijain2305/966/base -> origin/gh/anijain2305/966/base 2025-12-04T08:53:58.7448414Z * [new branch] gh/anijain2305/966/head -> origin/gh/anijain2305/966/head 2025-12-04T08:53:58.7448609Z * [new branch] gh/anijain2305/966/orig -> origin/gh/anijain2305/966/orig 2025-12-04T08:53:58.7448800Z * [new branch] gh/anijain2305/967/base -> origin/gh/anijain2305/967/base 2025-12-04T08:53:58.7449034Z * [new branch] gh/anijain2305/967/head -> origin/gh/anijain2305/967/head 2025-12-04T08:53:58.7449224Z * [new branch] gh/anijain2305/967/orig -> origin/gh/anijain2305/967/orig 2025-12-04T08:53:58.7449412Z * [new branch] gh/anijain2305/968/base -> origin/gh/anijain2305/968/base 2025-12-04T08:53:58.7449605Z * [new branch] gh/anijain2305/968/head -> origin/gh/anijain2305/968/head 2025-12-04T08:53:58.7449792Z * [new branch] gh/anijain2305/968/orig -> origin/gh/anijain2305/968/orig 2025-12-04T08:53:58.7449984Z * [new branch] gh/anijain2305/969/base -> origin/gh/anijain2305/969/base 2025-12-04T08:53:58.7450178Z * [new branch] gh/anijain2305/969/head -> origin/gh/anijain2305/969/head 2025-12-04T08:53:58.7450366Z * [new branch] gh/anijain2305/969/orig -> origin/gh/anijain2305/969/orig 2025-12-04T08:53:58.7450565Z * [new branch] gh/anijain2305/970/base -> origin/gh/anijain2305/970/base 2025-12-04T08:53:58.7450759Z * [new branch] gh/anijain2305/970/head -> origin/gh/anijain2305/970/head 2025-12-04T08:53:58.7450950Z * [new branch] gh/anijain2305/970/orig -> origin/gh/anijain2305/970/orig 2025-12-04T08:53:58.7451143Z * [new branch] gh/anjali411/216/base -> origin/gh/anjali411/216/base 2025-12-04T08:53:58.7451333Z * [new branch] gh/anjali411/216/head -> origin/gh/anjali411/216/head 2025-12-04T08:53:58.7451521Z * [new branch] gh/anjali411/216/orig -> origin/gh/anjali411/216/orig 2025-12-04T08:53:58.7451707Z * [new branch] gh/anshul-si/1/base -> origin/gh/anshul-si/1/base 2025-12-04T08:53:58.7451894Z * [new branch] gh/anshul-si/1/head -> origin/gh/anshul-si/1/head 2025-12-04T08:53:58.7452078Z * [new branch] gh/anshul-si/2/base -> origin/gh/anshul-si/2/base 2025-12-04T08:53:58.7452263Z * [new branch] gh/anshul-si/2/head -> origin/gh/anshul-si/2/head 2025-12-04T08:53:58.7452443Z * [new branch] gh/anshul-si/3/base -> origin/gh/anshul-si/3/base 2025-12-04T08:53:58.7452623Z * [new branch] gh/anshul-si/3/head -> origin/gh/anshul-si/3/head 2025-12-04T08:53:58.7452809Z * [new branch] gh/anshul-si/4/base -> origin/gh/anshul-si/4/base 2025-12-04T08:53:58.7452993Z * [new branch] gh/anshul-si/4/head -> origin/gh/anshul-si/4/head 2025-12-04T08:53:58.7453172Z * [new branch] gh/anshul-si/5/base -> origin/gh/anshul-si/5/base 2025-12-04T08:53:58.7453403Z * [new branch] gh/anshul-si/5/head -> origin/gh/anshul-si/5/head 2025-12-04T08:53:58.7453584Z * [new branch] gh/anshul-si/53/base -> origin/gh/anshul-si/53/base 2025-12-04T08:53:58.7453772Z * [new branch] gh/anshul-si/53/head -> origin/gh/anshul-si/53/head 2025-12-04T08:53:58.7453957Z * [new branch] gh/anshul-si/58/base -> origin/gh/anshul-si/58/base 2025-12-04T08:53:58.7454139Z * [new branch] gh/anshul-si/58/head -> origin/gh/anshul-si/58/head 2025-12-04T08:53:58.7454324Z * [new branch] gh/anshul-si/66/base -> origin/gh/anshul-si/66/base 2025-12-04T08:53:58.7454508Z * [new branch] gh/anshul-si/66/head -> origin/gh/anshul-si/66/head 2025-12-04T08:53:58.7454690Z * [new branch] gh/anshul-si/66/orig -> origin/gh/anshul-si/66/orig 2025-12-04T08:53:58.7454873Z * [new branch] gh/anshul-si/67/base -> origin/gh/anshul-si/67/base 2025-12-04T08:53:58.7455058Z * [new branch] gh/anshul-si/67/head -> origin/gh/anshul-si/67/head 2025-12-04T08:53:58.7455239Z * [new branch] gh/anshul-si/67/orig -> origin/gh/anshul-si/67/orig 2025-12-04T08:53:58.7455425Z * [new branch] gh/anshul-si/68/base -> origin/gh/anshul-si/68/base 2025-12-04T08:53:58.7455646Z * [new branch] gh/anshul-si/68/head -> origin/gh/anshul-si/68/head 2025-12-04T08:53:58.7455828Z * [new branch] gh/anshul-si/68/orig -> origin/gh/anshul-si/68/orig 2025-12-04T08:53:58.7456014Z * [new branch] gh/anshul-si/69/base -> origin/gh/anshul-si/69/base 2025-12-04T08:53:58.7456193Z * [new branch] gh/anshul-si/69/head -> origin/gh/anshul-si/69/head 2025-12-04T08:53:58.7456374Z * [new branch] gh/anshul-si/69/orig -> origin/gh/anshul-si/69/orig 2025-12-04T08:53:58.7456552Z * [new branch] gh/anshul-si/70/base -> origin/gh/anshul-si/70/base 2025-12-04T08:53:58.7456737Z * [new branch] gh/anshul-si/70/head -> origin/gh/anshul-si/70/head 2025-12-04T08:53:58.7456968Z * [new branch] gh/anshul-si/70/orig -> origin/gh/anshul-si/70/orig 2025-12-04T08:53:58.7457148Z * [new branch] gh/anshul-si/71/base -> origin/gh/anshul-si/71/base 2025-12-04T08:53:58.7457330Z * [new branch] gh/anshul-si/71/head -> origin/gh/anshul-si/71/head 2025-12-04T08:53:58.7457508Z * [new branch] gh/anshul-si/71/orig -> origin/gh/anshul-si/71/orig 2025-12-04T08:53:58.7457691Z * [new branch] gh/anshul-si/72/base -> origin/gh/anshul-si/72/base 2025-12-04T08:53:58.7457872Z * [new branch] gh/anshul-si/72/head -> origin/gh/anshul-si/72/head 2025-12-04T08:53:58.7458049Z * [new branch] gh/anshul-si/72/orig -> origin/gh/anshul-si/72/orig 2025-12-04T08:53:58.7458238Z * [new branch] gh/anshul-si/73/base -> origin/gh/anshul-si/73/base 2025-12-04T08:53:58.7458427Z * [new branch] gh/anshul-si/73/head -> origin/gh/anshul-si/73/head 2025-12-04T08:53:58.7458609Z * [new branch] gh/anshul-si/73/orig -> origin/gh/anshul-si/73/orig 2025-12-04T08:53:58.7458803Z * [new branch] gh/aorenste/132/base -> origin/gh/aorenste/132/base 2025-12-04T08:53:58.7458990Z * [new branch] gh/aorenste/132/head -> origin/gh/aorenste/132/head 2025-12-04T08:53:58.7459173Z * [new branch] gh/aorenste/134/base -> origin/gh/aorenste/134/base 2025-12-04T08:53:58.7459355Z * [new branch] gh/aorenste/134/head -> origin/gh/aorenste/134/head 2025-12-04T08:53:58.7459537Z * [new branch] gh/aorenste/134/orig -> origin/gh/aorenste/134/orig 2025-12-04T08:53:58.7459719Z * [new branch] gh/aorenste/139/base -> origin/gh/aorenste/139/base 2025-12-04T08:53:58.7459909Z * [new branch] gh/aorenste/139/head -> origin/gh/aorenste/139/head 2025-12-04T08:53:58.7460144Z * [new branch] gh/aorenste/139/orig -> origin/gh/aorenste/139/orig 2025-12-04T08:53:58.7460327Z * [new branch] gh/aorenste/141/base -> origin/gh/aorenste/141/base 2025-12-04T08:53:58.7460517Z * [new branch] gh/aorenste/141/head -> origin/gh/aorenste/141/head 2025-12-04T08:53:58.7460707Z * [new branch] gh/aorenste/145/base -> origin/gh/aorenste/145/base 2025-12-04T08:53:58.7460891Z * [new branch] gh/aorenste/145/head -> origin/gh/aorenste/145/head 2025-12-04T08:53:58.7461083Z * [new branch] gh/aorenste/145/orig -> origin/gh/aorenste/145/orig 2025-12-04T08:53:58.7461268Z * [new branch] gh/aorenste/146/base -> origin/gh/aorenste/146/base 2025-12-04T08:53:58.7461459Z * [new branch] gh/aorenste/146/head -> origin/gh/aorenste/146/head 2025-12-04T08:53:58.7461651Z * [new branch] gh/aorenste/146/orig -> origin/gh/aorenste/146/orig 2025-12-04T08:53:58.7461836Z * [new branch] gh/aorenste/147/base -> origin/gh/aorenste/147/base 2025-12-04T08:53:58.7462026Z * [new branch] gh/aorenste/147/head -> origin/gh/aorenste/147/head 2025-12-04T08:53:58.7462215Z * [new branch] gh/aorenste/147/orig -> origin/gh/aorenste/147/orig 2025-12-04T08:53:58.7462445Z * [new branch] gh/aorenste/148/base -> origin/gh/aorenste/148/base 2025-12-04T08:53:58.7462636Z * [new branch] gh/aorenste/148/head -> origin/gh/aorenste/148/head 2025-12-04T08:53:58.7462825Z * [new branch] gh/aorenste/148/orig -> origin/gh/aorenste/148/orig 2025-12-04T08:53:58.7463010Z * [new branch] gh/aorenste/149/base -> origin/gh/aorenste/149/base 2025-12-04T08:53:58.7463198Z * [new branch] gh/aorenste/149/head -> origin/gh/aorenste/149/head 2025-12-04T08:53:58.7463390Z * [new branch] gh/aorenste/149/orig -> origin/gh/aorenste/149/orig 2025-12-04T08:53:58.7463574Z * [new branch] gh/aorenste/150/base -> origin/gh/aorenste/150/base 2025-12-04T08:53:58.7463761Z * [new branch] gh/aorenste/150/head -> origin/gh/aorenste/150/head 2025-12-04T08:53:58.7463949Z * [new branch] gh/aorenste/150/orig -> origin/gh/aorenste/150/orig 2025-12-04T08:53:58.7464136Z * [new branch] gh/aorenste/151/base -> origin/gh/aorenste/151/base 2025-12-04T08:53:58.7464326Z * [new branch] gh/aorenste/151/head -> origin/gh/aorenste/151/head 2025-12-04T08:53:58.7464514Z * [new branch] gh/aorenste/151/orig -> origin/gh/aorenste/151/orig 2025-12-04T08:53:58.7464700Z * [new branch] gh/aorenste/152/base -> origin/gh/aorenste/152/base 2025-12-04T08:53:58.7464892Z * [new branch] gh/aorenste/152/head -> origin/gh/aorenste/152/head 2025-12-04T08:53:58.7465083Z * [new branch] gh/aorenste/152/orig -> origin/gh/aorenste/152/orig 2025-12-04T08:53:58.7465268Z * [new branch] gh/aorenste/153/base -> origin/gh/aorenste/153/base 2025-12-04T08:53:58.7465458Z * [new branch] gh/aorenste/153/head -> origin/gh/aorenste/153/head 2025-12-04T08:53:58.7465644Z * [new branch] gh/aorenste/153/orig -> origin/gh/aorenste/153/orig 2025-12-04T08:53:58.7465835Z * [new branch] gh/aorenste/154/base -> origin/gh/aorenste/154/base 2025-12-04T08:53:58.7466026Z * [new branch] gh/aorenste/154/head -> origin/gh/aorenste/154/head 2025-12-04T08:53:58.7466212Z * [new branch] gh/aorenste/154/orig -> origin/gh/aorenste/154/orig 2025-12-04T08:53:58.7466401Z * [new branch] gh/aorenste/155/base -> origin/gh/aorenste/155/base 2025-12-04T08:53:58.7466591Z * [new branch] gh/aorenste/155/head -> origin/gh/aorenste/155/head 2025-12-04T08:53:58.7466853Z * [new branch] gh/aorenste/155/orig -> origin/gh/aorenste/155/orig 2025-12-04T08:53:58.7467045Z * [new branch] gh/aorenste/156/base -> origin/gh/aorenste/156/base 2025-12-04T08:53:58.7467234Z * [new branch] gh/aorenste/156/head -> origin/gh/aorenste/156/head 2025-12-04T08:53:58.7467419Z * [new branch] gh/aorenste/156/orig -> origin/gh/aorenste/156/orig 2025-12-04T08:53:58.7467611Z * [new branch] gh/aorenste/157/base -> origin/gh/aorenste/157/base 2025-12-04T08:53:58.7467801Z * [new branch] gh/aorenste/157/head -> origin/gh/aorenste/157/head 2025-12-04T08:53:58.7467986Z * [new branch] gh/aorenste/157/orig -> origin/gh/aorenste/157/orig 2025-12-04T08:53:58.7468175Z * [new branch] gh/aorenste/158/base -> origin/gh/aorenste/158/base 2025-12-04T08:53:58.7468366Z * [new branch] gh/aorenste/158/head -> origin/gh/aorenste/158/head 2025-12-04T08:53:58.7468553Z * [new branch] gh/aorenste/158/orig -> origin/gh/aorenste/158/orig 2025-12-04T08:53:58.7468740Z * [new branch] gh/aorenste/159/base -> origin/gh/aorenste/159/base 2025-12-04T08:53:58.7468931Z * [new branch] gh/aorenste/159/head -> origin/gh/aorenste/159/head 2025-12-04T08:53:58.7469157Z * [new branch] gh/aorenste/159/orig -> origin/gh/aorenste/159/orig 2025-12-04T08:53:58.7469358Z * [new branch] gh/avikchaudhuri/1/base -> origin/gh/avikchaudhuri/1/base 2025-12-04T08:53:58.7469560Z * [new branch] gh/avikchaudhuri/1/head -> origin/gh/avikchaudhuri/1/head 2025-12-04T08:53:58.7469763Z * [new branch] gh/avikchaudhuri/2/base -> origin/gh/avikchaudhuri/2/base 2025-12-04T08:53:58.7469964Z * [new branch] gh/avikchaudhuri/2/head -> origin/gh/avikchaudhuri/2/head 2025-12-04T08:53:58.7470162Z * [new branch] gh/avikchaudhuri/2/orig -> origin/gh/avikchaudhuri/2/orig 2025-12-04T08:53:58.7470362Z * [new branch] gh/bdhirsh/666/base -> origin/gh/bdhirsh/666/base 2025-12-04T08:53:58.7470551Z * [new branch] gh/bdhirsh/666/head -> origin/gh/bdhirsh/666/head 2025-12-04T08:53:58.7470735Z * [new branch] gh/bdhirsh/666/orig -> origin/gh/bdhirsh/666/orig 2025-12-04T08:53:58.7470923Z * [new branch] gh/bdhirsh/668/base -> origin/gh/bdhirsh/668/base 2025-12-04T08:53:58.7471112Z * [new branch] gh/bdhirsh/668/head -> origin/gh/bdhirsh/668/head 2025-12-04T08:53:58.7471296Z * [new branch] gh/bdhirsh/668/orig -> origin/gh/bdhirsh/668/orig 2025-12-04T08:53:58.7471482Z * [new branch] gh/bdhirsh/669/base -> origin/gh/bdhirsh/669/base 2025-12-04T08:53:58.7471668Z * [new branch] gh/bdhirsh/669/head -> origin/gh/bdhirsh/669/head 2025-12-04T08:53:58.7471850Z * [new branch] gh/bdhirsh/669/orig -> origin/gh/bdhirsh/669/orig 2025-12-04T08:53:58.7472037Z * [new branch] gh/bdhirsh/670/base -> origin/gh/bdhirsh/670/base 2025-12-04T08:53:58.7472223Z * [new branch] gh/bdhirsh/670/head -> origin/gh/bdhirsh/670/head 2025-12-04T08:53:58.7472405Z * [new branch] gh/bdhirsh/670/orig -> origin/gh/bdhirsh/670/orig 2025-12-04T08:53:58.7472593Z * [new branch] gh/bdhirsh/672/base -> origin/gh/bdhirsh/672/base 2025-12-04T08:53:58.7472779Z * [new branch] gh/bdhirsh/672/head -> origin/gh/bdhirsh/672/head 2025-12-04T08:53:58.7472962Z * [new branch] gh/bdhirsh/672/orig -> origin/gh/bdhirsh/672/orig 2025-12-04T08:53:58.7473147Z * [new branch] gh/bdhirsh/675/base -> origin/gh/bdhirsh/675/base 2025-12-04T08:53:58.7473327Z * [new branch] gh/bdhirsh/675/head -> origin/gh/bdhirsh/675/head 2025-12-04T08:53:58.7473513Z * [new branch] gh/bdhirsh/675/orig -> origin/gh/bdhirsh/675/orig 2025-12-04T08:53:58.7473752Z * [new branch] gh/bdhirsh/676/base -> origin/gh/bdhirsh/676/base 2025-12-04T08:53:58.7473934Z * [new branch] gh/bdhirsh/676/head -> origin/gh/bdhirsh/676/head 2025-12-04T08:53:58.7474121Z * [new branch] gh/bdhirsh/676/orig -> origin/gh/bdhirsh/676/orig 2025-12-04T08:53:58.7474310Z * [new branch] gh/bdhirsh/677/base -> origin/gh/bdhirsh/677/base 2025-12-04T08:53:58.7474493Z * [new branch] gh/bdhirsh/677/head -> origin/gh/bdhirsh/677/head 2025-12-04T08:53:58.7474571Z * [new branch] gh/bdhirsh/677/orig -> origin/gh/bdhirsh/677/orig 2025-12-04T08:53:58.7474644Z * [new branch] gh/bdhirsh/678/base -> origin/gh/bdhirsh/678/base 2025-12-04T08:53:58.7474716Z * [new branch] gh/bdhirsh/678/head -> origin/gh/bdhirsh/678/head 2025-12-04T08:53:58.7474793Z * [new branch] gh/bdhirsh/678/orig -> origin/gh/bdhirsh/678/orig 2025-12-04T08:53:58.7474870Z * [new branch] gh/bdhirsh/679/base -> origin/gh/bdhirsh/679/base 2025-12-04T08:53:58.7474942Z * [new branch] gh/bdhirsh/679/head -> origin/gh/bdhirsh/679/head 2025-12-04T08:53:58.7475020Z * [new branch] gh/bdhirsh/679/orig -> origin/gh/bdhirsh/679/orig 2025-12-04T08:53:58.7475123Z * [new branch] gh/bdhirsh/680/base -> origin/gh/bdhirsh/680/base 2025-12-04T08:53:58.7475200Z * [new branch] gh/bdhirsh/680/head -> origin/gh/bdhirsh/680/head 2025-12-04T08:53:58.7475272Z * [new branch] gh/bdhirsh/680/orig -> origin/gh/bdhirsh/680/orig 2025-12-04T08:53:58.7475344Z * [new branch] gh/bdhirsh/681/base -> origin/gh/bdhirsh/681/base 2025-12-04T08:53:58.7475420Z * [new branch] gh/bdhirsh/681/head -> origin/gh/bdhirsh/681/head 2025-12-04T08:53:58.7475492Z * [new branch] gh/bdhirsh/681/orig -> origin/gh/bdhirsh/681/orig 2025-12-04T08:53:58.7475590Z * [new branch] gh/benjaminglass1/101/base -> origin/gh/benjaminglass1/101/base 2025-12-04T08:53:58.7475687Z * [new branch] gh/benjaminglass1/101/head -> origin/gh/benjaminglass1/101/head 2025-12-04T08:53:58.7475777Z * [new branch] gh/benjaminglass1/101/orig -> origin/gh/benjaminglass1/101/orig 2025-12-04T08:53:58.7475869Z * [new branch] gh/benjaminglass1/102/base -> origin/gh/benjaminglass1/102/base 2025-12-04T08:53:58.7475965Z * [new branch] gh/benjaminglass1/102/head -> origin/gh/benjaminglass1/102/head 2025-12-04T08:53:58.7476054Z * [new branch] gh/benjaminglass1/102/orig -> origin/gh/benjaminglass1/102/orig 2025-12-04T08:53:58.7476143Z * [new branch] gh/benjaminglass1/106/base -> origin/gh/benjaminglass1/106/base 2025-12-04T08:53:58.7476237Z * [new branch] gh/benjaminglass1/106/head -> origin/gh/benjaminglass1/106/head 2025-12-04T08:53:58.7476327Z * [new branch] gh/benjaminglass1/106/orig -> origin/gh/benjaminglass1/106/orig 2025-12-04T08:53:58.7476415Z * [new branch] gh/benjaminglass1/107/base -> origin/gh/benjaminglass1/107/base 2025-12-04T08:53:58.7476508Z * [new branch] gh/benjaminglass1/107/head -> origin/gh/benjaminglass1/107/head 2025-12-04T08:53:58.7476599Z * [new branch] gh/benjaminglass1/107/orig -> origin/gh/benjaminglass1/107/orig 2025-12-04T08:53:58.7476687Z * [new branch] gh/benjaminglass1/108/base -> origin/gh/benjaminglass1/108/base 2025-12-04T08:53:58.7476825Z * [new branch] gh/benjaminglass1/108/head -> origin/gh/benjaminglass1/108/head 2025-12-04T08:53:58.7476916Z * [new branch] gh/benjaminglass1/108/orig -> origin/gh/benjaminglass1/108/orig 2025-12-04T08:53:58.7477010Z * [new branch] gh/benjaminglass1/109/base -> origin/gh/benjaminglass1/109/base 2025-12-04T08:53:58.7477140Z * [new branch] gh/benjaminglass1/109/head -> origin/gh/benjaminglass1/109/head 2025-12-04T08:53:58.7477229Z * [new branch] gh/benjaminglass1/109/orig -> origin/gh/benjaminglass1/109/orig 2025-12-04T08:53:58.7477321Z * [new branch] gh/benjaminglass1/97/base -> origin/gh/benjaminglass1/97/base 2025-12-04T08:53:58.7477410Z * [new branch] gh/benjaminglass1/97/head -> origin/gh/benjaminglass1/97/head 2025-12-04T08:53:58.7477498Z * [new branch] gh/benjaminglass1/97/orig -> origin/gh/benjaminglass1/97/orig 2025-12-04T08:53:58.7477584Z * [new branch] gh/bobrenjc93/570/base -> origin/gh/bobrenjc93/570/base 2025-12-04T08:53:58.7477665Z * [new branch] gh/bobrenjc93/570/head -> origin/gh/bobrenjc93/570/head 2025-12-04T08:53:58.7477744Z * [new branch] gh/bobrenjc93/570/orig -> origin/gh/bobrenjc93/570/orig 2025-12-04T08:53:58.7477825Z * [new branch] gh/bobrenjc93/604/base -> origin/gh/bobrenjc93/604/base 2025-12-04T08:53:58.7477904Z * [new branch] gh/bobrenjc93/604/head -> origin/gh/bobrenjc93/604/head 2025-12-04T08:53:58.7477978Z * [new branch] gh/bobrenjc93/604/orig -> origin/gh/bobrenjc93/604/orig 2025-12-04T08:53:58.7478059Z * [new branch] gh/bobrenjc93/638/base -> origin/gh/bobrenjc93/638/base 2025-12-04T08:53:58.7478186Z * [new branch] gh/bobrenjc93/638/head -> origin/gh/bobrenjc93/638/head 2025-12-04T08:53:58.7478263Z * [new branch] gh/bobrenjc93/638/orig -> origin/gh/bobrenjc93/638/orig 2025-12-04T08:53:58.7478343Z * [new branch] gh/bobrenjc93/653/base -> origin/gh/bobrenjc93/653/base 2025-12-04T08:53:58.7478418Z * [new branch] gh/bobrenjc93/653/head -> origin/gh/bobrenjc93/653/head 2025-12-04T08:53:58.7478498Z * [new branch] gh/bobrenjc93/653/orig -> origin/gh/bobrenjc93/653/orig 2025-12-04T08:53:58.7478574Z * [new branch] gh/bobrenjc93/654/base -> origin/gh/bobrenjc93/654/base 2025-12-04T08:53:58.7478652Z * [new branch] gh/bobrenjc93/654/head -> origin/gh/bobrenjc93/654/head 2025-12-04T08:53:58.7478733Z * [new branch] gh/bobrenjc93/654/orig -> origin/gh/bobrenjc93/654/orig 2025-12-04T08:53:58.7478809Z * [new branch] gh/bobrenjc93/657/base -> origin/gh/bobrenjc93/657/base 2025-12-04T08:53:58.7478887Z * [new branch] gh/bobrenjc93/657/head -> origin/gh/bobrenjc93/657/head 2025-12-04T08:53:58.7478967Z * [new branch] gh/bobrenjc93/657/orig -> origin/gh/bobrenjc93/657/orig 2025-12-04T08:53:58.7479043Z * [new branch] gh/bobrenjc93/672/base -> origin/gh/bobrenjc93/672/base 2025-12-04T08:53:58.7479121Z * [new branch] gh/bobrenjc93/672/head -> origin/gh/bobrenjc93/672/head 2025-12-04T08:53:58.7479201Z * [new branch] gh/bobrenjc93/672/orig -> origin/gh/bobrenjc93/672/orig 2025-12-04T08:53:58.7479279Z * [new branch] gh/bobrenjc93/679/base -> origin/gh/bobrenjc93/679/base 2025-12-04T08:53:58.7479355Z * [new branch] gh/bobrenjc93/679/head -> origin/gh/bobrenjc93/679/head 2025-12-04T08:53:58.7479435Z * [new branch] gh/bobrenjc93/679/orig -> origin/gh/bobrenjc93/679/orig 2025-12-04T08:53:58.7479511Z * [new branch] gh/bobrenjc93/680/base -> origin/gh/bobrenjc93/680/base 2025-12-04T08:53:58.7479590Z * [new branch] gh/bobrenjc93/680/head -> origin/gh/bobrenjc93/680/head 2025-12-04T08:53:58.7479670Z * [new branch] gh/bobrenjc93/680/orig -> origin/gh/bobrenjc93/680/orig 2025-12-04T08:53:58.7479746Z * [new branch] gh/bobrenjc93/681/base -> origin/gh/bobrenjc93/681/base 2025-12-04T08:53:58.7479821Z * [new branch] gh/bobrenjc93/681/head -> origin/gh/bobrenjc93/681/head 2025-12-04T08:53:58.7479901Z * [new branch] gh/bobrenjc93/681/orig -> origin/gh/bobrenjc93/681/orig 2025-12-04T08:53:58.7480012Z * [new branch] gh/bobrenjc93/682/base -> origin/gh/bobrenjc93/682/base 2025-12-04T08:53:58.7480089Z * [new branch] gh/bobrenjc93/682/head -> origin/gh/bobrenjc93/682/head 2025-12-04T08:53:58.7480168Z * [new branch] gh/bobrenjc93/682/orig -> origin/gh/bobrenjc93/682/orig 2025-12-04T08:53:58.7480245Z * [new branch] gh/bobrenjc93/683/base -> origin/gh/bobrenjc93/683/base 2025-12-04T08:53:58.7480325Z * [new branch] gh/bobrenjc93/683/head -> origin/gh/bobrenjc93/683/head 2025-12-04T08:53:58.7480400Z * [new branch] gh/bobrenjc93/683/orig -> origin/gh/bobrenjc93/683/orig 2025-12-04T08:53:58.7480476Z * [new branch] gh/bobrenjc93/684/base -> origin/gh/bobrenjc93/684/base 2025-12-04T08:53:58.7480556Z * [new branch] gh/bobrenjc93/684/head -> origin/gh/bobrenjc93/684/head 2025-12-04T08:53:58.7480631Z * [new branch] gh/bobrenjc93/684/orig -> origin/gh/bobrenjc93/684/orig 2025-12-04T08:53:58.7480709Z * [new branch] gh/bobrenjc93/685/base -> origin/gh/bobrenjc93/685/base 2025-12-04T08:53:58.7480791Z * [new branch] gh/bobrenjc93/685/head -> origin/gh/bobrenjc93/685/head 2025-12-04T08:53:58.7480866Z * [new branch] gh/bobrenjc93/685/orig -> origin/gh/bobrenjc93/685/orig 2025-12-04T08:53:58.7480966Z * [new branch] gh/bobrenjc93/686/base -> origin/gh/bobrenjc93/686/base 2025-12-04T08:53:58.7481047Z * [new branch] gh/bobrenjc93/686/head -> origin/gh/bobrenjc93/686/head 2025-12-04T08:53:58.7481123Z * [new branch] gh/bobrenjc93/686/orig -> origin/gh/bobrenjc93/686/orig 2025-12-04T08:53:58.7481198Z * [new branch] gh/bobrenjc93/687/base -> origin/gh/bobrenjc93/687/base 2025-12-04T08:53:58.7481278Z * [new branch] gh/bobrenjc93/687/head -> origin/gh/bobrenjc93/687/head 2025-12-04T08:53:58.7481354Z * [new branch] gh/bobrenjc93/687/orig -> origin/gh/bobrenjc93/687/orig 2025-12-04T08:53:58.7481431Z * [new branch] gh/bobrenjc93/688/base -> origin/gh/bobrenjc93/688/base 2025-12-04T08:53:58.7481510Z * [new branch] gh/bobrenjc93/688/head -> origin/gh/bobrenjc93/688/head 2025-12-04T08:53:58.7481585Z * [new branch] gh/bobrenjc93/688/orig -> origin/gh/bobrenjc93/688/orig 2025-12-04T08:53:58.7481661Z * [new branch] gh/bobrenjc93/689/base -> origin/gh/bobrenjc93/689/base 2025-12-04T08:53:58.7481741Z * [new branch] gh/bobrenjc93/689/head -> origin/gh/bobrenjc93/689/head 2025-12-04T08:53:58.7481817Z * [new branch] gh/bobrenjc93/689/orig -> origin/gh/bobrenjc93/689/orig 2025-12-04T08:53:58.7481897Z * [new branch] gh/bobrenjc93/690/base -> origin/gh/bobrenjc93/690/base 2025-12-04T08:53:58.7481973Z * [new branch] gh/bobrenjc93/690/head -> origin/gh/bobrenjc93/690/head 2025-12-04T08:53:58.7482050Z * [new branch] gh/bobrenjc93/690/orig -> origin/gh/bobrenjc93/690/orig 2025-12-04T08:53:58.7482130Z * [new branch] gh/bobrenjc93/691/base -> origin/gh/bobrenjc93/691/base 2025-12-04T08:53:58.7482206Z * [new branch] gh/bobrenjc93/691/head -> origin/gh/bobrenjc93/691/head 2025-12-04T08:53:58.7482282Z * [new branch] gh/bobrenjc93/691/orig -> origin/gh/bobrenjc93/691/orig 2025-12-04T08:53:58.7482363Z * [new branch] gh/bobrenjc93/692/base -> origin/gh/bobrenjc93/692/base 2025-12-04T08:53:58.7482440Z * [new branch] gh/bobrenjc93/692/head -> origin/gh/bobrenjc93/692/head 2025-12-04T08:53:58.7482516Z * [new branch] gh/bobrenjc93/692/orig -> origin/gh/bobrenjc93/692/orig 2025-12-04T08:53:58.7482595Z * [new branch] gh/bobrenjc93/693/base -> origin/gh/bobrenjc93/693/base 2025-12-04T08:53:58.7482670Z * [new branch] gh/bobrenjc93/693/head -> origin/gh/bobrenjc93/693/head 2025-12-04T08:53:58.7482775Z * [new branch] gh/bobrenjc93/693/orig -> origin/gh/bobrenjc93/693/orig 2025-12-04T08:53:58.7482856Z * [new branch] gh/bobrenjc93/694/base -> origin/gh/bobrenjc93/694/base 2025-12-04T08:53:58.7482933Z * [new branch] gh/bobrenjc93/694/head -> origin/gh/bobrenjc93/694/head 2025-12-04T08:53:58.7483010Z * [new branch] gh/bobrenjc93/694/orig -> origin/gh/bobrenjc93/694/orig 2025-12-04T08:53:58.7483090Z * [new branch] gh/bobrenjc93/695/base -> origin/gh/bobrenjc93/695/base 2025-12-04T08:53:58.7483167Z * [new branch] gh/bobrenjc93/695/head -> origin/gh/bobrenjc93/695/head 2025-12-04T08:53:58.7483242Z * [new branch] gh/bobrenjc93/695/orig -> origin/gh/bobrenjc93/695/orig 2025-12-04T08:53:58.7483318Z * [new branch] gh/c00w/23/base -> origin/gh/c00w/23/base 2025-12-04T08:53:58.7483388Z * [new branch] gh/c00w/23/head -> origin/gh/c00w/23/head 2025-12-04T08:53:58.7483459Z * [new branch] gh/c00w/53/base -> origin/gh/c00w/53/base 2025-12-04T08:53:58.7483527Z * [new branch] gh/c00w/53/head -> origin/gh/c00w/53/head 2025-12-04T08:53:58.7483593Z * [new branch] gh/c00w/53/orig -> origin/gh/c00w/53/orig 2025-12-04T08:53:58.7483690Z * [new branch] gh/c00w/54/base -> origin/gh/c00w/54/base 2025-12-04T08:53:58.7483756Z * [new branch] gh/c00w/54/head -> origin/gh/c00w/54/head 2025-12-04T08:53:58.7483824Z * [new branch] gh/c00w/54/orig -> origin/gh/c00w/54/orig 2025-12-04T08:53:58.7483895Z * [new branch] gh/c00w/56/base -> origin/gh/c00w/56/base 2025-12-04T08:53:58.7483962Z * [new branch] gh/c00w/56/head -> origin/gh/c00w/56/head 2025-12-04T08:53:58.7484029Z * [new branch] gh/c00w/56/orig -> origin/gh/c00w/56/orig 2025-12-04T08:53:58.7484101Z * [new branch] gh/c00w/57/base -> origin/gh/c00w/57/base 2025-12-04T08:53:58.7484168Z * [new branch] gh/c00w/57/head -> origin/gh/c00w/57/head 2025-12-04T08:53:58.7484234Z * [new branch] gh/c00w/57/orig -> origin/gh/c00w/57/orig 2025-12-04T08:53:58.7484304Z * [new branch] gh/c00w/58/base -> origin/gh/c00w/58/base 2025-12-04T08:53:58.7484372Z * [new branch] gh/c00w/58/head -> origin/gh/c00w/58/head 2025-12-04T08:53:58.7484438Z * [new branch] gh/c00w/58/orig -> origin/gh/c00w/58/orig 2025-12-04T08:53:58.7484518Z * [new branch] gh/clee2000/1/base -> origin/gh/clee2000/1/base 2025-12-04T08:53:58.7484593Z * [new branch] gh/clee2000/1/head -> origin/gh/clee2000/1/head 2025-12-04T08:53:58.7484667Z * [new branch] gh/clee2000/1/orig -> origin/gh/clee2000/1/orig 2025-12-04T08:53:58.7484754Z * [new branch] gh/coconutruben/1/base -> origin/gh/coconutruben/1/base 2025-12-04T08:53:58.7484834Z * [new branch] gh/coconutruben/1/head -> origin/gh/coconutruben/1/head 2025-12-04T08:53:58.7484918Z * [new branch] gh/coconutruben/55/base -> origin/gh/coconutruben/55/base 2025-12-04T08:53:58.7485004Z * [new branch] gh/coconutruben/55/head -> origin/gh/coconutruben/55/head 2025-12-04T08:53:58.7485085Z * [new branch] gh/coconutruben/55/orig -> origin/gh/coconutruben/55/orig 2025-12-04T08:53:58.7485165Z * [new branch] gh/coconutruben/57/base -> origin/gh/coconutruben/57/base 2025-12-04T08:53:58.7485249Z * [new branch] gh/coconutruben/57/head -> origin/gh/coconutruben/57/head 2025-12-04T08:53:58.7485328Z * [new branch] gh/coconutruben/57/orig -> origin/gh/coconutruben/57/orig 2025-12-04T08:53:58.7485411Z * [new branch] gh/coconutruben/70/base -> origin/gh/coconutruben/70/base 2025-12-04T08:53:58.7485529Z * [new branch] gh/coconutruben/70/head -> origin/gh/coconutruben/70/head 2025-12-04T08:53:58.7485608Z * [new branch] gh/coconutruben/70/orig -> origin/gh/coconutruben/70/orig 2025-12-04T08:53:58.7485692Z * [new branch] gh/coconutruben/71/base -> origin/gh/coconutruben/71/base 2025-12-04T08:53:58.7485773Z * [new branch] gh/coconutruben/71/head -> origin/gh/coconutruben/71/head 2025-12-04T08:53:58.7485852Z * [new branch] gh/coconutruben/71/orig -> origin/gh/coconutruben/71/orig 2025-12-04T08:53:58.7485937Z * [new branch] gh/coconutruben/72/base -> origin/gh/coconutruben/72/base 2025-12-04T08:53:58.7486016Z * [new branch] gh/coconutruben/72/head -> origin/gh/coconutruben/72/head 2025-12-04T08:53:58.7486095Z * [new branch] gh/coconutruben/72/orig -> origin/gh/coconutruben/72/orig 2025-12-04T08:53:58.7486178Z * [new branch] gh/coconutruben/73/base -> origin/gh/coconutruben/73/base 2025-12-04T08:53:58.7486257Z * [new branch] gh/coconutruben/73/head -> origin/gh/coconutruben/73/head 2025-12-04T08:53:58.7486336Z * [new branch] gh/coconutruben/73/orig -> origin/gh/coconutruben/73/orig 2025-12-04T08:53:58.7486419Z * [new branch] gh/coconutruben/74/base -> origin/gh/coconutruben/74/base 2025-12-04T08:53:58.7486531Z * [new branch] gh/coconutruben/74/head -> origin/gh/coconutruben/74/head 2025-12-04T08:53:58.7486613Z * [new branch] gh/coconutruben/74/orig -> origin/gh/coconutruben/74/orig 2025-12-04T08:53:58.7486696Z * [new branch] gh/coconutruben/79/base -> origin/gh/coconutruben/79/base 2025-12-04T08:53:58.7486830Z * [new branch] gh/coconutruben/79/head -> origin/gh/coconutruben/79/head 2025-12-04T08:53:58.7486914Z * [new branch] gh/coconutruben/79/orig -> origin/gh/coconutruben/79/orig 2025-12-04T08:53:58.7486998Z * [new branch] gh/coconutruben/80/base -> origin/gh/coconutruben/80/base 2025-12-04T08:53:58.7487077Z * [new branch] gh/coconutruben/80/head -> origin/gh/coconutruben/80/head 2025-12-04T08:53:58.7487162Z * [new branch] gh/coconutruben/80/orig -> origin/gh/coconutruben/80/orig 2025-12-04T08:53:58.7487241Z * [new branch] gh/coconutruben/82/base -> origin/gh/coconutruben/82/base 2025-12-04T08:53:58.7487320Z * [new branch] gh/coconutruben/82/head -> origin/gh/coconutruben/82/head 2025-12-04T08:53:58.7487405Z * [new branch] gh/coconutruben/82/orig -> origin/gh/coconutruben/82/orig 2025-12-04T08:53:58.7487484Z * [new branch] gh/coconutruben/83/base -> origin/gh/coconutruben/83/base 2025-12-04T08:53:58.7487563Z * [new branch] gh/coconutruben/83/head -> origin/gh/coconutruben/83/head 2025-12-04T08:53:58.7487649Z * [new branch] gh/coconutruben/83/orig -> origin/gh/coconutruben/83/orig 2025-12-04T08:53:58.7487729Z * [new branch] gh/coconutruben/84/base -> origin/gh/coconutruben/84/base 2025-12-04T08:53:58.7487810Z * [new branch] gh/coconutruben/84/head -> origin/gh/coconutruben/84/head 2025-12-04T08:53:58.7487893Z * [new branch] gh/coconutruben/84/orig -> origin/gh/coconutruben/84/orig 2025-12-04T08:53:58.7487973Z * [new branch] gh/coconutruben/85/base -> origin/gh/coconutruben/85/base 2025-12-04T08:53:58.7488052Z * [new branch] gh/coconutruben/85/head -> origin/gh/coconutruben/85/head 2025-12-04T08:53:58.7488135Z * [new branch] gh/coconutruben/85/orig -> origin/gh/coconutruben/85/orig 2025-12-04T08:53:58.7488213Z * [new branch] gh/coconutruben/86/base -> origin/gh/coconutruben/86/base 2025-12-04T08:53:58.7488293Z * [new branch] gh/coconutruben/86/head -> origin/gh/coconutruben/86/head 2025-12-04T08:53:58.7488376Z * [new branch] gh/coconutruben/86/orig -> origin/gh/coconutruben/86/orig 2025-12-04T08:53:58.7488500Z * [new branch] gh/colinchan15/1/base -> origin/gh/colinchan15/1/base 2025-12-04T08:53:58.7488578Z * [new branch] gh/colinchan15/1/head -> origin/gh/colinchan15/1/head 2025-12-04T08:53:58.7488659Z * [new branch] gh/colinchan15/2/base -> origin/gh/colinchan15/2/base 2025-12-04T08:53:58.7488738Z * [new branch] gh/colinchan15/2/head -> origin/gh/colinchan15/2/head 2025-12-04T08:53:58.7488815Z * [new branch] gh/colinchan15/3/base -> origin/gh/colinchan15/3/base 2025-12-04T08:53:58.7488895Z * [new branch] gh/colinchan15/3/head -> origin/gh/colinchan15/3/head 2025-12-04T08:53:58.7488971Z * [new branch] gh/colinchan15/6/base -> origin/gh/colinchan15/6/base 2025-12-04T08:53:58.7489050Z * [new branch] gh/colinchan15/6/head -> origin/gh/colinchan15/6/head 2025-12-04T08:53:58.7489119Z * [new branch] gh/d4l3k/1/base -> origin/gh/d4l3k/1/base 2025-12-04T08:53:58.7489189Z * [new branch] gh/d4l3k/1/head -> origin/gh/d4l3k/1/head 2025-12-04T08:53:58.7489259Z * [new branch] gh/d4l3k/2/base -> origin/gh/d4l3k/2/base 2025-12-04T08:53:58.7489326Z * [new branch] gh/d4l3k/2/head -> origin/gh/d4l3k/2/head 2025-12-04T08:53:58.7489443Z * [new branch] gh/d4l3k/2/orig -> origin/gh/d4l3k/2/orig 2025-12-04T08:53:58.7489515Z * [new branch] gh/d4l3k/3/base -> origin/gh/d4l3k/3/base 2025-12-04T08:53:58.7489582Z * [new branch] gh/d4l3k/3/head -> origin/gh/d4l3k/3/head 2025-12-04T08:53:58.7489649Z * [new branch] gh/d4l3k/3/orig -> origin/gh/d4l3k/3/orig 2025-12-04T08:53:58.7489719Z * [new branch] gh/d4l3k/4/base -> origin/gh/d4l3k/4/base 2025-12-04T08:53:58.7489785Z * [new branch] gh/d4l3k/4/head -> origin/gh/d4l3k/4/head 2025-12-04T08:53:58.7489852Z * [new branch] gh/d4l3k/4/orig -> origin/gh/d4l3k/4/orig 2025-12-04T08:53:58.7489921Z * [new branch] gh/d4l3k/5/base -> origin/gh/d4l3k/5/base 2025-12-04T08:53:58.7489987Z * [new branch] gh/d4l3k/5/orig -> origin/gh/d4l3k/5/orig 2025-12-04T08:53:58.7490078Z * [new branch] gh/davidberard98/392/base -> origin/gh/davidberard98/392/base 2025-12-04T08:53:58.7490172Z * [new branch] gh/davidberard98/392/head -> origin/gh/davidberard98/392/head 2025-12-04T08:53:58.7490257Z * [new branch] gh/davidberard98/392/orig -> origin/gh/davidberard98/392/orig 2025-12-04T08:53:58.7490343Z * [new branch] gh/davidberard98/399/base -> origin/gh/davidberard98/399/base 2025-12-04T08:53:58.7490433Z * [new branch] gh/davidberard98/399/head -> origin/gh/davidberard98/399/head 2025-12-04T08:53:58.7490515Z * [new branch] gh/davidberard98/399/orig -> origin/gh/davidberard98/399/orig 2025-12-04T08:53:58.7490594Z * [new branch] gh/desertfire/605/base -> origin/gh/desertfire/605/base 2025-12-04T08:53:58.7490674Z * [new branch] gh/desertfire/605/head -> origin/gh/desertfire/605/head 2025-12-04T08:53:58.7490752Z * [new branch] gh/desertfire/605/orig -> origin/gh/desertfire/605/orig 2025-12-04T08:53:58.7490835Z * [new branch] gh/desertfire/606/base -> origin/gh/desertfire/606/base 2025-12-04T08:53:58.7490911Z * [new branch] gh/desertfire/606/head -> origin/gh/desertfire/606/head 2025-12-04T08:53:58.7490986Z * [new branch] gh/desertfire/606/orig -> origin/gh/desertfire/606/orig 2025-12-04T08:53:58.7491063Z * [new branch] gh/desertfire/607/base -> origin/gh/desertfire/607/base 2025-12-04T08:53:58.7491139Z * [new branch] gh/desertfire/607/head -> origin/gh/desertfire/607/head 2025-12-04T08:53:58.7491242Z * [new branch] gh/desertfire/607/orig -> origin/gh/desertfire/607/orig 2025-12-04T08:53:58.7491320Z * [new branch] gh/desertfire/608/base -> origin/gh/desertfire/608/base 2025-12-04T08:53:58.7491396Z * [new branch] gh/desertfire/608/head -> origin/gh/desertfire/608/head 2025-12-04T08:53:58.7491472Z * [new branch] gh/desertfire/608/orig -> origin/gh/desertfire/608/orig 2025-12-04T08:53:58.7491552Z * [new branch] gh/desertfire/609/base -> origin/gh/desertfire/609/base 2025-12-04T08:53:58.7491627Z * [new branch] gh/desertfire/609/head -> origin/gh/desertfire/609/head 2025-12-04T08:53:58.7491701Z * [new branch] gh/desertfire/609/orig -> origin/gh/desertfire/609/orig 2025-12-04T08:53:58.7491780Z * [new branch] gh/desertfire/610/base -> origin/gh/desertfire/610/base 2025-12-04T08:53:58.7491855Z * [new branch] gh/desertfire/610/head -> origin/gh/desertfire/610/head 2025-12-04T08:53:58.7491933Z * [new branch] gh/desertfire/610/orig -> origin/gh/desertfire/610/orig 2025-12-04T08:53:58.7492012Z * [new branch] gh/desertfire/611/base -> origin/gh/desertfire/611/base 2025-12-04T08:53:58.7492088Z * [new branch] gh/desertfire/611/head -> origin/gh/desertfire/611/head 2025-12-04T08:53:58.7492165Z * [new branch] gh/desertfire/611/orig -> origin/gh/desertfire/611/orig 2025-12-04T08:53:58.7492277Z * [new branch] gh/desertfire/612/base -> origin/gh/desertfire/612/base 2025-12-04T08:53:58.7492352Z * [new branch] gh/desertfire/612/head -> origin/gh/desertfire/612/head 2025-12-04T08:53:58.7492429Z * [new branch] gh/desertfire/612/orig -> origin/gh/desertfire/612/orig 2025-12-04T08:53:58.7492504Z * [new branch] gh/desertfire/613/base -> origin/gh/desertfire/613/base 2025-12-04T08:53:58.7492579Z * [new branch] gh/desertfire/613/head -> origin/gh/desertfire/613/head 2025-12-04T08:53:58.7492660Z * [new branch] gh/desertfire/613/orig -> origin/gh/desertfire/613/orig 2025-12-04T08:53:58.7492738Z * [new branch] gh/desertfire/614/base -> origin/gh/desertfire/614/base 2025-12-04T08:53:58.7492812Z * [new branch] gh/desertfire/614/head -> origin/gh/desertfire/614/head 2025-12-04T08:53:58.7492891Z * [new branch] gh/desertfire/614/orig -> origin/gh/desertfire/614/orig 2025-12-04T08:53:58.7492967Z * [new branch] gh/desertfire/615/base -> origin/gh/desertfire/615/base 2025-12-04T08:53:58.7493042Z * [new branch] gh/desertfire/615/head -> origin/gh/desertfire/615/head 2025-12-04T08:53:58.7493120Z * [new branch] gh/desertfire/615/orig -> origin/gh/desertfire/615/orig 2025-12-04T08:53:58.7493195Z * [new branch] gh/desertfire/616/base -> origin/gh/desertfire/616/base 2025-12-04T08:53:58.7493269Z * [new branch] gh/desertfire/616/head -> origin/gh/desertfire/616/head 2025-12-04T08:53:58.7493351Z * [new branch] gh/desertfire/616/orig -> origin/gh/desertfire/616/orig 2025-12-04T08:53:58.7493428Z * [new branch] gh/desertfire/617/base -> origin/gh/desertfire/617/base 2025-12-04T08:53:58.7493504Z * [new branch] gh/desertfire/617/head -> origin/gh/desertfire/617/head 2025-12-04T08:53:58.7493582Z * [new branch] gh/desertfire/617/orig -> origin/gh/desertfire/617/orig 2025-12-04T08:53:58.7493656Z * [new branch] gh/dharakk/1/base -> origin/gh/dharakk/1/base 2025-12-04T08:53:58.7493727Z * [new branch] gh/dharakk/1/head -> origin/gh/dharakk/1/head 2025-12-04T08:53:58.7493803Z * [new branch] gh/drisspg/170/base -> origin/gh/drisspg/170/base 2025-12-04T08:53:58.7493876Z * [new branch] gh/drisspg/170/head -> origin/gh/drisspg/170/head 2025-12-04T08:53:58.7493954Z * [new branch] gh/drisspg/170/orig -> origin/gh/drisspg/170/orig 2025-12-04T08:53:58.7494057Z * [new branch] gh/drisspg/182/base -> origin/gh/drisspg/182/base 2025-12-04T08:53:58.7494128Z * [new branch] gh/drisspg/182/head -> origin/gh/drisspg/182/head 2025-12-04T08:53:58.7494202Z * [new branch] gh/drisspg/183/base -> origin/gh/drisspg/183/base 2025-12-04T08:53:58.7494274Z * [new branch] gh/drisspg/183/head -> origin/gh/drisspg/183/head 2025-12-04T08:53:58.7494345Z * [new branch] gh/drisspg/184/base -> origin/gh/drisspg/184/base 2025-12-04T08:53:58.7494419Z * [new branch] gh/drisspg/184/head -> origin/gh/drisspg/184/head 2025-12-04T08:53:58.7494490Z * [new branch] gh/drisspg/185/base -> origin/gh/drisspg/185/base 2025-12-04T08:53:58.7494562Z * [new branch] gh/drisspg/185/head -> origin/gh/drisspg/185/head 2025-12-04T08:53:58.7494636Z * [new branch] gh/drisspg/194/base -> origin/gh/drisspg/194/base 2025-12-04T08:53:58.7494708Z * [new branch] gh/drisspg/194/head -> origin/gh/drisspg/194/head 2025-12-04T08:53:58.7494780Z * [new branch] gh/drisspg/194/orig -> origin/gh/drisspg/194/orig 2025-12-04T08:53:58.7494853Z * [new branch] gh/drisspg/200/base -> origin/gh/drisspg/200/base 2025-12-04T08:53:58.7494949Z * [new branch] gh/drisspg/200/head -> origin/gh/drisspg/200/head 2025-12-04T08:53:58.7495021Z * [new branch] gh/drisspg/200/orig -> origin/gh/drisspg/200/orig 2025-12-04T08:53:58.7495096Z * [new branch] gh/drisspg/218/base -> origin/gh/drisspg/218/base 2025-12-04T08:53:58.7495168Z * [new branch] gh/drisspg/218/head -> origin/gh/drisspg/218/head 2025-12-04T08:53:58.7495239Z * [new branch] gh/drisspg/218/orig -> origin/gh/drisspg/218/orig 2025-12-04T08:53:58.7495314Z * [new branch] gh/drisspg/219/base -> origin/gh/drisspg/219/base 2025-12-04T08:53:58.7495387Z * [new branch] gh/drisspg/219/head -> origin/gh/drisspg/219/head 2025-12-04T08:53:58.7495458Z * [new branch] gh/drisspg/219/orig -> origin/gh/drisspg/219/orig 2025-12-04T08:53:58.7495531Z * [new branch] gh/drisspg/220/base -> origin/gh/drisspg/220/base 2025-12-04T08:53:58.7495603Z * [new branch] gh/drisspg/220/head -> origin/gh/drisspg/220/head 2025-12-04T08:53:58.7495675Z * [new branch] gh/drisspg/220/orig -> origin/gh/drisspg/220/orig 2025-12-04T08:53:58.7495747Z * [new branch] gh/drisspg/221/base -> origin/gh/drisspg/221/base 2025-12-04T08:53:58.7495818Z * [new branch] gh/drisspg/221/head -> origin/gh/drisspg/221/head 2025-12-04T08:53:58.7495894Z * [new branch] gh/drisspg/221/orig -> origin/gh/drisspg/221/orig 2025-12-04T08:53:58.7495965Z * [new branch] gh/drisspg/222/base -> origin/gh/drisspg/222/base 2025-12-04T08:53:58.7496038Z * [new branch] gh/drisspg/222/head -> origin/gh/drisspg/222/head 2025-12-04T08:53:58.7496110Z * [new branch] gh/drisspg/222/orig -> origin/gh/drisspg/222/orig 2025-12-04T08:53:58.7496181Z * [new branch] gh/drisspg/223/base -> origin/gh/drisspg/223/base 2025-12-04T08:53:58.7496251Z * [new branch] gh/drisspg/223/head -> origin/gh/drisspg/223/head 2025-12-04T08:53:58.7496324Z * [new branch] gh/drisspg/223/orig -> origin/gh/drisspg/223/orig 2025-12-04T08:53:58.7496394Z * [new branch] gh/drisspg/224/base -> origin/gh/drisspg/224/base 2025-12-04T08:53:58.7496465Z * [new branch] gh/drisspg/224/head -> origin/gh/drisspg/224/head 2025-12-04T08:53:58.7496540Z * [new branch] gh/drisspg/224/orig -> origin/gh/drisspg/224/orig 2025-12-04T08:53:58.7496616Z * [new branch] gh/drisspg/225/base -> origin/gh/drisspg/225/base 2025-12-04T08:53:58.7496721Z * [new branch] gh/drisspg/225/head -> origin/gh/drisspg/225/head 2025-12-04T08:53:58.7496838Z * [new branch] gh/drisspg/225/orig -> origin/gh/drisspg/225/orig 2025-12-04T08:53:58.7496911Z * [new branch] gh/drisspg/226/base -> origin/gh/drisspg/226/base 2025-12-04T08:53:58.7496982Z * [new branch] gh/drisspg/226/head -> origin/gh/drisspg/226/head 2025-12-04T08:53:58.7497055Z * [new branch] gh/drisspg/226/orig -> origin/gh/drisspg/226/orig 2025-12-04T08:53:58.7497132Z * [new branch] gh/drisspg/227/base -> origin/gh/drisspg/227/base 2025-12-04T08:53:58.7497210Z * [new branch] gh/drisspg/227/head -> origin/gh/drisspg/227/head 2025-12-04T08:53:58.7497418Z * [new branch] gh/drisspg/227/orig -> origin/gh/drisspg/227/orig 2025-12-04T08:53:58.7497618Z * [new branch] gh/drisspg/228/base -> origin/gh/drisspg/228/base 2025-12-04T08:53:58.7497809Z * [new branch] gh/drisspg/228/head -> origin/gh/drisspg/228/head 2025-12-04T08:53:58.7497996Z * [new branch] gh/drisspg/228/orig -> origin/gh/drisspg/228/orig 2025-12-04T08:53:58.7498179Z * [new branch] gh/drisspg/229/base -> origin/gh/drisspg/229/base 2025-12-04T08:53:58.7498401Z * [new branch] gh/drisspg/229/head -> origin/gh/drisspg/229/head 2025-12-04T08:53:58.7498585Z * [new branch] gh/drisspg/229/orig -> origin/gh/drisspg/229/orig 2025-12-04T08:53:58.7498769Z * [new branch] gh/drisspg/230/base -> origin/gh/drisspg/230/base 2025-12-04T08:53:58.7498961Z * [new branch] gh/drisspg/230/head -> origin/gh/drisspg/230/head 2025-12-04T08:53:58.7499145Z * [new branch] gh/drisspg/230/orig -> origin/gh/drisspg/230/orig 2025-12-04T08:53:58.7499337Z * [new branch] gh/dsjohns2/1/base -> origin/gh/dsjohns2/1/base 2025-12-04T08:53:58.7499533Z * [new branch] gh/dsjohns2/1/head -> origin/gh/dsjohns2/1/head 2025-12-04T08:53:58.7499730Z * [new branch] gh/dzmitry-huba/1/base -> origin/gh/dzmitry-huba/1/base 2025-12-04T08:53:58.7499937Z * [new branch] gh/dzmitry-huba/1/head -> origin/gh/dzmitry-huba/1/head 2025-12-04T08:53:58.7500134Z * [new branch] gh/dzmitry-huba/12/base -> origin/gh/dzmitry-huba/12/base 2025-12-04T08:53:58.7500330Z * [new branch] gh/dzmitry-huba/12/head -> origin/gh/dzmitry-huba/12/head 2025-12-04T08:53:58.7500531Z * [new branch] gh/dzmitry-huba/12/orig -> origin/gh/dzmitry-huba/12/orig 2025-12-04T08:53:58.7500721Z * [new branch] gh/dzmitry-huba/13/base -> origin/gh/dzmitry-huba/13/base 2025-12-04T08:53:58.7500918Z * [new branch] gh/dzmitry-huba/13/head -> origin/gh/dzmitry-huba/13/head 2025-12-04T08:53:58.7501112Z * [new branch] gh/dzmitry-huba/13/orig -> origin/gh/dzmitry-huba/13/orig 2025-12-04T08:53:58.7501304Z * [new branch] gh/dzmitry-huba/14/base -> origin/gh/dzmitry-huba/14/base 2025-12-04T08:53:58.7501496Z * [new branch] gh/dzmitry-huba/14/head -> origin/gh/dzmitry-huba/14/head 2025-12-04T08:53:58.7501688Z * [new branch] gh/dzmitry-huba/14/orig -> origin/gh/dzmitry-huba/14/orig 2025-12-04T08:53:58.7501882Z * [new branch] gh/dzmitry-huba/15/base -> origin/gh/dzmitry-huba/15/base 2025-12-04T08:53:58.7502074Z * [new branch] gh/dzmitry-huba/15/head -> origin/gh/dzmitry-huba/15/head 2025-12-04T08:53:58.7502265Z * [new branch] gh/dzmitry-huba/15/orig -> origin/gh/dzmitry-huba/15/orig 2025-12-04T08:53:58.7502463Z * [new branch] gh/dzmitry-huba/16/base -> origin/gh/dzmitry-huba/16/base 2025-12-04T08:53:58.7502670Z * [new branch] gh/dzmitry-huba/16/head -> origin/gh/dzmitry-huba/16/head 2025-12-04T08:53:58.7502919Z * [new branch] gh/dzmitry-huba/16/orig -> origin/gh/dzmitry-huba/16/orig 2025-12-04T08:53:58.7503126Z * [new branch] gh/dzmitry-huba/17/base -> origin/gh/dzmitry-huba/17/base 2025-12-04T08:53:58.7503322Z * [new branch] gh/dzmitry-huba/17/head -> origin/gh/dzmitry-huba/17/head 2025-12-04T08:53:58.7503532Z * [new branch] gh/dzmitry-huba/17/orig -> origin/gh/dzmitry-huba/17/orig 2025-12-04T08:53:58.7503737Z * [new branch] gh/dzmitry-huba/2/base -> origin/gh/dzmitry-huba/2/base 2025-12-04T08:53:58.7503945Z * [new branch] gh/dzmitry-huba/2/head -> origin/gh/dzmitry-huba/2/head 2025-12-04T08:53:58.7504140Z * [new branch] gh/dzmitry-huba/3/base -> origin/gh/dzmitry-huba/3/base 2025-12-04T08:53:58.7504331Z * [new branch] gh/dzmitry-huba/3/head -> origin/gh/dzmitry-huba/3/head 2025-12-04T08:53:58.7504571Z * [new branch] gh/eellison/808/base -> origin/gh/eellison/808/base 2025-12-04T08:53:58.7504841Z * [new branch] gh/eellison/808/head -> origin/gh/eellison/808/head 2025-12-04T08:53:58.7505057Z * [new branch] gh/eellison/808/orig -> origin/gh/eellison/808/orig 2025-12-04T08:53:58.7505243Z * [new branch] gh/eellison/822/base -> origin/gh/eellison/822/base 2025-12-04T08:53:58.7505499Z * [new branch] gh/eellison/822/head -> origin/gh/eellison/822/head 2025-12-04T08:53:58.7505695Z * [new branch] gh/eellison/822/orig -> origin/gh/eellison/822/orig 2025-12-04T08:53:58.7505876Z * [new branch] gh/eellison/823/base -> origin/gh/eellison/823/base 2025-12-04T08:53:58.7506059Z * [new branch] gh/eellison/823/head -> origin/gh/eellison/823/head 2025-12-04T08:53:58.7506244Z * [new branch] gh/eellison/823/orig -> origin/gh/eellison/823/orig 2025-12-04T08:53:58.7506427Z * [new branch] gh/eellison/862/base -> origin/gh/eellison/862/base 2025-12-04T08:53:58.7506615Z * [new branch] gh/eellison/862/head -> origin/gh/eellison/862/head 2025-12-04T08:53:58.7506872Z * [new branch] gh/eellison/862/orig -> origin/gh/eellison/862/orig 2025-12-04T08:53:58.7507057Z * [new branch] gh/eellison/863/base -> origin/gh/eellison/863/base 2025-12-04T08:53:58.7507249Z * [new branch] gh/eellison/863/head -> origin/gh/eellison/863/head 2025-12-04T08:53:58.7507431Z * [new branch] gh/eellison/863/orig -> origin/gh/eellison/863/orig 2025-12-04T08:53:58.7507615Z * [new branch] gh/eellison/864/base -> origin/gh/eellison/864/base 2025-12-04T08:53:58.7507797Z * [new branch] gh/eellison/864/head -> origin/gh/eellison/864/head 2025-12-04T08:53:58.7507979Z * [new branch] gh/eellison/864/orig -> origin/gh/eellison/864/orig 2025-12-04T08:53:58.7508167Z * [new branch] gh/eellison/865/base -> origin/gh/eellison/865/base 2025-12-04T08:53:58.7508351Z * [new branch] gh/eellison/865/head -> origin/gh/eellison/865/head 2025-12-04T08:53:58.7508533Z * [new branch] gh/eellison/865/orig -> origin/gh/eellison/865/orig 2025-12-04T08:53:58.7508718Z * [new branch] gh/eellison/866/base -> origin/gh/eellison/866/base 2025-12-04T08:53:58.7508903Z * [new branch] gh/eellison/866/head -> origin/gh/eellison/866/head 2025-12-04T08:53:58.7509088Z * [new branch] gh/eellison/866/orig -> origin/gh/eellison/866/orig 2025-12-04T08:53:58.7509272Z * [new branch] gh/eellison/867/base -> origin/gh/eellison/867/base 2025-12-04T08:53:58.7509456Z * [new branch] gh/eellison/867/head -> origin/gh/eellison/867/head 2025-12-04T08:53:58.7509640Z * [new branch] gh/eellison/867/orig -> origin/gh/eellison/867/orig 2025-12-04T08:53:58.7509879Z * [new branch] gh/eellison/868/base -> origin/gh/eellison/868/base 2025-12-04T08:53:58.7510066Z * [new branch] gh/eellison/868/head -> origin/gh/eellison/868/head 2025-12-04T08:53:58.7510252Z * [new branch] gh/eellison/868/orig -> origin/gh/eellison/868/orig 2025-12-04T08:53:58.7510438Z * [new branch] gh/eellison/869/base -> origin/gh/eellison/869/base 2025-12-04T08:53:58.7510637Z * [new branch] gh/eellison/869/head -> origin/gh/eellison/869/head 2025-12-04T08:53:58.7510823Z * [new branch] gh/eellison/869/orig -> origin/gh/eellison/869/orig 2025-12-04T08:53:58.7511007Z * [new branch] gh/eellison/870/base -> origin/gh/eellison/870/base 2025-12-04T08:53:58.7511186Z * [new branch] gh/eellison/870/head -> origin/gh/eellison/870/head 2025-12-04T08:53:58.7511369Z * [new branch] gh/eellison/870/orig -> origin/gh/eellison/870/orig 2025-12-04T08:53:58.7511553Z * [new branch] gh/eellison/871/base -> origin/gh/eellison/871/base 2025-12-04T08:53:58.7511732Z * [new branch] gh/eellison/871/head -> origin/gh/eellison/871/head 2025-12-04T08:53:58.7511913Z * [new branch] gh/eellison/871/orig -> origin/gh/eellison/871/orig 2025-12-04T08:53:58.7512093Z * [new branch] gh/eellison/872/base -> origin/gh/eellison/872/base 2025-12-04T08:53:58.7512316Z * [new branch] gh/eellison/872/head -> origin/gh/eellison/872/head 2025-12-04T08:53:58.7512498Z * [new branch] gh/eellison/872/orig -> origin/gh/eellison/872/orig 2025-12-04T08:53:58.7512677Z * [new branch] gh/eellison/873/base -> origin/gh/eellison/873/base 2025-12-04T08:53:58.7512855Z * [new branch] gh/eellison/873/head -> origin/gh/eellison/873/head 2025-12-04T08:53:58.7513035Z * [new branch] gh/eellison/873/orig -> origin/gh/eellison/873/orig 2025-12-04T08:53:58.7513217Z * [new branch] gh/eellison/874/base -> origin/gh/eellison/874/base 2025-12-04T08:53:58.7513395Z * [new branch] gh/eellison/874/head -> origin/gh/eellison/874/head 2025-12-04T08:53:58.7513574Z * [new branch] gh/eellison/874/orig -> origin/gh/eellison/874/orig 2025-12-04T08:53:58.7513753Z * [new branch] gh/eellison/875/base -> origin/gh/eellison/875/base 2025-12-04T08:53:58.7513935Z * [new branch] gh/eellison/875/head -> origin/gh/eellison/875/head 2025-12-04T08:53:58.7514114Z * [new branch] gh/eellison/875/orig -> origin/gh/eellison/875/orig 2025-12-04T08:53:58.7514293Z * [new branch] gh/eellison/876/base -> origin/gh/eellison/876/base 2025-12-04T08:53:58.7514476Z * [new branch] gh/eellison/876/head -> origin/gh/eellison/876/head 2025-12-04T08:53:58.7514656Z * [new branch] gh/eellison/876/orig -> origin/gh/eellison/876/orig 2025-12-04T08:53:58.7514836Z * [new branch] gh/eellison/877/base -> origin/gh/eellison/877/base 2025-12-04T08:53:58.7515016Z * [new branch] gh/eellison/877/head -> origin/gh/eellison/877/head 2025-12-04T08:53:58.7515196Z * [new branch] gh/eellison/877/orig -> origin/gh/eellison/877/orig 2025-12-04T08:53:58.7515378Z * [new branch] gh/eellison/878/base -> origin/gh/eellison/878/base 2025-12-04T08:53:58.7515557Z * [new branch] gh/eellison/878/head -> origin/gh/eellison/878/head 2025-12-04T08:53:58.7515737Z * [new branch] gh/eellison/878/orig -> origin/gh/eellison/878/orig 2025-12-04T08:53:58.7515915Z * [new branch] gh/eellison/879/base -> origin/gh/eellison/879/base 2025-12-04T08:53:58.7516097Z * [new branch] gh/eellison/879/head -> origin/gh/eellison/879/head 2025-12-04T08:53:58.7516280Z * [new branch] gh/eellison/879/orig -> origin/gh/eellison/879/orig 2025-12-04T08:53:58.7516492Z * [new branch] gh/eellison/880/base -> origin/gh/eellison/880/base 2025-12-04T08:53:58.7516671Z * [new branch] gh/eellison/880/head -> origin/gh/eellison/880/head 2025-12-04T08:53:58.7516891Z * [new branch] gh/eellison/880/orig -> origin/gh/eellison/880/orig 2025-12-04T08:53:58.7517071Z * [new branch] gh/eellison/881/base -> origin/gh/eellison/881/base 2025-12-04T08:53:58.7517251Z * [new branch] gh/eellison/881/head -> origin/gh/eellison/881/head 2025-12-04T08:53:58.7517434Z * [new branch] gh/eellison/881/orig -> origin/gh/eellison/881/orig 2025-12-04T08:53:58.7517619Z * [new branch] gh/eellison/882/base -> origin/gh/eellison/882/base 2025-12-04T08:53:58.7517804Z * [new branch] gh/eellison/882/head -> origin/gh/eellison/882/head 2025-12-04T08:53:58.7517986Z * [new branch] gh/eellison/882/orig -> origin/gh/eellison/882/orig 2025-12-04T08:53:58.7518173Z * [new branch] gh/eellison/883/base -> origin/gh/eellison/883/base 2025-12-04T08:53:58.7518358Z * [new branch] gh/eellison/883/head -> origin/gh/eellison/883/head 2025-12-04T08:53:58.7518543Z * [new branch] gh/eellison/883/orig -> origin/gh/eellison/883/orig 2025-12-04T08:53:58.7518772Z * [new branch] gh/eellison/884/base -> origin/gh/eellison/884/base 2025-12-04T08:53:58.7518955Z * [new branch] gh/eellison/884/head -> origin/gh/eellison/884/head 2025-12-04T08:53:58.7519133Z * [new branch] gh/eellison/884/orig -> origin/gh/eellison/884/orig 2025-12-04T08:53:58.7519313Z * [new branch] gh/etaf/147/base -> origin/gh/etaf/147/base 2025-12-04T08:53:58.7519485Z * [new branch] gh/etaf/147/head -> origin/gh/etaf/147/head 2025-12-04T08:53:58.7519655Z * [new branch] gh/etaf/154/base -> origin/gh/etaf/154/base 2025-12-04T08:53:58.7519829Z * [new branch] gh/etaf/154/head -> origin/gh/etaf/154/head 2025-12-04T08:53:58.7519999Z * [new branch] gh/etaf/154/orig -> origin/gh/etaf/154/orig 2025-12-04T08:53:58.7520167Z * [new branch] gh/etaf/156/base -> origin/gh/etaf/156/base 2025-12-04T08:53:58.7520337Z * [new branch] gh/etaf/156/head -> origin/gh/etaf/156/head 2025-12-04T08:53:58.7538743Z * [new branch] gh/etaf/156/orig -> origin/gh/etaf/156/orig 2025-12-04T08:53:58.7538975Z * [new branch] gh/etaf/157/base -> origin/gh/etaf/157/base 2025-12-04T08:53:58.7539157Z * [new branch] gh/etaf/157/head -> origin/gh/etaf/157/head 2025-12-04T08:53:58.7539334Z * [new branch] gh/etaf/157/orig -> origin/gh/etaf/157/orig 2025-12-04T08:53:58.7539539Z * [new branch] gh/etaf/158/base -> origin/gh/etaf/158/base 2025-12-04T08:53:58.7539732Z * [new branch] gh/etaf/158/head -> origin/gh/etaf/158/head 2025-12-04T08:53:58.7539900Z * [new branch] gh/etaf/158/orig -> origin/gh/etaf/158/orig 2025-12-04T08:53:58.7540071Z * [new branch] gh/etaf/159/base -> origin/gh/etaf/159/base 2025-12-04T08:53:58.7540243Z * [new branch] gh/etaf/159/head -> origin/gh/etaf/159/head 2025-12-04T08:53:58.7540423Z * [new branch] gh/etaf/159/orig -> origin/gh/etaf/159/orig 2025-12-04T08:53:58.7540596Z * [new branch] gh/etaf/160/base -> origin/gh/etaf/160/base 2025-12-04T08:53:58.7540772Z * [new branch] gh/etaf/160/head -> origin/gh/etaf/160/head 2025-12-04T08:53:58.7540964Z * [new branch] gh/etaf/160/orig -> origin/gh/etaf/160/orig 2025-12-04T08:53:58.7541139Z * [new branch] gh/etaf/161/base -> origin/gh/etaf/161/base 2025-12-04T08:53:58.7541383Z * [new branch] gh/etaf/161/head -> origin/gh/etaf/161/head 2025-12-04T08:53:58.7541552Z * [new branch] gh/etaf/161/orig -> origin/gh/etaf/161/orig 2025-12-04T08:53:58.7541730Z * [new branch] gh/etaf/166/base -> origin/gh/etaf/166/base 2025-12-04T08:53:58.7541897Z * [new branch] gh/etaf/166/head -> origin/gh/etaf/166/head 2025-12-04T08:53:58.7542077Z * [new branch] gh/etaf/166/orig -> origin/gh/etaf/166/orig 2025-12-04T08:53:58.7542247Z * [new branch] gh/etaf/167/base -> origin/gh/etaf/167/base 2025-12-04T08:53:58.7542420Z * [new branch] gh/etaf/167/head -> origin/gh/etaf/167/head 2025-12-04T08:53:58.7542594Z * [new branch] gh/etaf/167/orig -> origin/gh/etaf/167/orig 2025-12-04T08:53:58.7542763Z * [new branch] gh/etaf/168/base -> origin/gh/etaf/168/base 2025-12-04T08:53:58.7542934Z * [new branch] gh/etaf/168/head -> origin/gh/etaf/168/head 2025-12-04T08:53:58.7543105Z * [new branch] gh/etaf/168/orig -> origin/gh/etaf/168/orig 2025-12-04T08:53:58.7543272Z * [new branch] gh/etaf/172/base -> origin/gh/etaf/172/base 2025-12-04T08:53:58.7543440Z * [new branch] gh/etaf/172/head -> origin/gh/etaf/172/head 2025-12-04T08:53:58.7543658Z * [new branch] gh/etaf/172/orig -> origin/gh/etaf/172/orig 2025-12-04T08:53:58.7543826Z * [new branch] gh/etaf/173/base -> origin/gh/etaf/173/base 2025-12-04T08:53:58.7543996Z * [new branch] gh/etaf/173/head -> origin/gh/etaf/173/head 2025-12-04T08:53:58.7544163Z * [new branch] gh/etaf/173/orig -> origin/gh/etaf/173/orig 2025-12-04T08:53:58.7544331Z * [new branch] gh/etaf/174/base -> origin/gh/etaf/174/base 2025-12-04T08:53:58.7544499Z * [new branch] gh/etaf/174/head -> origin/gh/etaf/174/head 2025-12-04T08:53:58.7544667Z * [new branch] gh/etaf/175/base -> origin/gh/etaf/175/base 2025-12-04T08:53:58.7544835Z * [new branch] gh/etaf/175/head -> origin/gh/etaf/175/head 2025-12-04T08:53:58.7545004Z * [new branch] gh/etaf/175/orig -> origin/gh/etaf/175/orig 2025-12-04T08:53:58.7545173Z * [new branch] gh/etaf/176/base -> origin/gh/etaf/176/base 2025-12-04T08:53:58.7545342Z * [new branch] gh/etaf/176/head -> origin/gh/etaf/176/head 2025-12-04T08:53:58.7545511Z * [new branch] gh/etaf/176/orig -> origin/gh/etaf/176/orig 2025-12-04T08:53:58.7545677Z * [new branch] gh/etaf/177/base -> origin/gh/etaf/177/base 2025-12-04T08:53:58.7545845Z * [new branch] gh/etaf/177/head -> origin/gh/etaf/177/head 2025-12-04T08:53:58.7546013Z * [new branch] gh/etaf/177/orig -> origin/gh/etaf/177/orig 2025-12-04T08:53:58.7546185Z * [new branch] gh/etaf/178/base -> origin/gh/etaf/178/base 2025-12-04T08:53:58.7546354Z * [new branch] gh/etaf/178/head -> origin/gh/etaf/178/head 2025-12-04T08:53:58.7546523Z * [new branch] gh/etaf/178/orig -> origin/gh/etaf/178/orig 2025-12-04T08:53:58.7546692Z * [new branch] gh/etaf/179/base -> origin/gh/etaf/179/base 2025-12-04T08:53:58.7546914Z * [new branch] gh/etaf/179/head -> origin/gh/etaf/179/head 2025-12-04T08:53:58.7547082Z * [new branch] gh/etaf/179/orig -> origin/gh/etaf/179/orig 2025-12-04T08:53:58.7547249Z * [new branch] gh/etaf/180/base -> origin/gh/etaf/180/base 2025-12-04T08:53:58.7547417Z * [new branch] gh/etaf/180/head -> origin/gh/etaf/180/head 2025-12-04T08:53:58.7547585Z * [new branch] gh/etaf/180/orig -> origin/gh/etaf/180/orig 2025-12-04T08:53:58.7547822Z * [new branch] gh/exclamaforte/1/base -> origin/gh/exclamaforte/1/base 2025-12-04T08:53:58.7548019Z * [new branch] gh/exclamaforte/1/head -> origin/gh/exclamaforte/1/head 2025-12-04T08:53:58.7548210Z * [new branch] gh/exclamaforte/2/base -> origin/gh/exclamaforte/2/base 2025-12-04T08:53:58.7548399Z * [new branch] gh/exclamaforte/2/head -> origin/gh/exclamaforte/2/head 2025-12-04T08:53:58.7548591Z * [new branch] gh/exclamaforte/3/base -> origin/gh/exclamaforte/3/base 2025-12-04T08:53:58.7548780Z * [new branch] gh/exclamaforte/3/head -> origin/gh/exclamaforte/3/head 2025-12-04T08:53:58.7548970Z * [new branch] gh/exclamaforte/4/base -> origin/gh/exclamaforte/4/base 2025-12-04T08:53:58.7549162Z * [new branch] gh/exclamaforte/4/head -> origin/gh/exclamaforte/4/head 2025-12-04T08:53:58.7549350Z * [new branch] gh/ezyang/2374/base -> origin/gh/ezyang/2374/base 2025-12-04T08:53:58.7549536Z * [new branch] gh/ezyang/2374/head -> origin/gh/ezyang/2374/head 2025-12-04T08:53:58.7549718Z * [new branch] gh/ezyang/2374/orig -> origin/gh/ezyang/2374/orig 2025-12-04T08:53:58.7549895Z * [new branch] gh/ezyang/2973/base -> origin/gh/ezyang/2973/base 2025-12-04T08:53:58.7550112Z * [new branch] gh/ezyang/2973/head -> origin/gh/ezyang/2973/head 2025-12-04T08:53:58.7550289Z * [new branch] gh/ezyang/2973/orig -> origin/gh/ezyang/2973/orig 2025-12-04T08:53:58.7550465Z * [new branch] gh/ezyang/2974/base -> origin/gh/ezyang/2974/base 2025-12-04T08:53:58.7550641Z * [new branch] gh/ezyang/2974/head -> origin/gh/ezyang/2974/head 2025-12-04T08:53:58.7550818Z * [new branch] gh/ezyang/2974/orig -> origin/gh/ezyang/2974/orig 2025-12-04T08:53:58.7550995Z * [new branch] gh/ezyang/3131/base -> origin/gh/ezyang/3131/base 2025-12-04T08:53:58.7551173Z * [new branch] gh/ezyang/3131/head -> origin/gh/ezyang/3131/head 2025-12-04T08:53:58.7551349Z * [new branch] gh/ezyang/3131/orig -> origin/gh/ezyang/3131/orig 2025-12-04T08:53:58.7551528Z * [new branch] gh/ezyang/3139/base -> origin/gh/ezyang/3139/base 2025-12-04T08:53:58.7551707Z * [new branch] gh/ezyang/3139/head -> origin/gh/ezyang/3139/head 2025-12-04T08:53:58.7551882Z * [new branch] gh/ezyang/3139/orig -> origin/gh/ezyang/3139/orig 2025-12-04T08:53:58.7552061Z * [new branch] gh/ezyang/3140/base -> origin/gh/ezyang/3140/base 2025-12-04T08:53:58.7552239Z * [new branch] gh/ezyang/3140/head -> origin/gh/ezyang/3140/head 2025-12-04T08:53:58.7552415Z * [new branch] gh/ezyang/3140/orig -> origin/gh/ezyang/3140/orig 2025-12-04T08:53:58.7552595Z * [new branch] gh/ezyang/3143/base -> origin/gh/ezyang/3143/base 2025-12-04T08:53:58.7552774Z * [new branch] gh/ezyang/3143/head -> origin/gh/ezyang/3143/head 2025-12-04T08:53:58.7552955Z * [new branch] gh/ezyang/3143/orig -> origin/gh/ezyang/3143/orig 2025-12-04T08:53:58.7553131Z * [new branch] gh/ezyang/3144/base -> origin/gh/ezyang/3144/base 2025-12-04T08:53:58.7553310Z * [new branch] gh/ezyang/3144/head -> origin/gh/ezyang/3144/head 2025-12-04T08:53:58.7553487Z * [new branch] gh/ezyang/3144/orig -> origin/gh/ezyang/3144/orig 2025-12-04T08:53:58.7553664Z * [new branch] gh/ezyang/3167/base -> origin/gh/ezyang/3167/base 2025-12-04T08:53:58.7553841Z * [new branch] gh/ezyang/3167/head -> origin/gh/ezyang/3167/head 2025-12-04T08:53:58.7554018Z * [new branch] gh/ezyang/3167/orig -> origin/gh/ezyang/3167/orig 2025-12-04T08:53:58.7554191Z * [new branch] gh/ezyang/3173/base -> origin/gh/ezyang/3173/base 2025-12-04T08:53:58.7554396Z * [new branch] gh/ezyang/3173/head -> origin/gh/ezyang/3173/head 2025-12-04T08:53:58.7554573Z * [new branch] gh/ezyang/3173/orig -> origin/gh/ezyang/3173/orig 2025-12-04T08:53:58.7554749Z * [new branch] gh/ezyang/3175/base -> origin/gh/ezyang/3175/base 2025-12-04T08:53:58.7554925Z * [new branch] gh/ezyang/3175/head -> origin/gh/ezyang/3175/head 2025-12-04T08:53:58.7555102Z * [new branch] gh/ezyang/3175/orig -> origin/gh/ezyang/3175/orig 2025-12-04T08:53:58.7555280Z * [new branch] gh/ezyang/3182/base -> origin/gh/ezyang/3182/base 2025-12-04T08:53:58.7555456Z * [new branch] gh/ezyang/3182/head -> origin/gh/ezyang/3182/head 2025-12-04T08:53:58.7555637Z * [new branch] gh/ezyang/3182/orig -> origin/gh/ezyang/3182/orig 2025-12-04T08:53:58.7555814Z * [new branch] gh/ezyang/3185/base -> origin/gh/ezyang/3185/base 2025-12-04T08:53:58.7555992Z * [new branch] gh/ezyang/3185/head -> origin/gh/ezyang/3185/head 2025-12-04T08:53:58.7556168Z * [new branch] gh/ezyang/3185/orig -> origin/gh/ezyang/3185/orig 2025-12-04T08:53:58.7556347Z * [new branch] gh/ezyang/3189/base -> origin/gh/ezyang/3189/base 2025-12-04T08:53:58.7556552Z * [new branch] gh/ezyang/3189/head -> origin/gh/ezyang/3189/head 2025-12-04T08:53:58.7556730Z * [new branch] gh/ezyang/3189/orig -> origin/gh/ezyang/3189/orig 2025-12-04T08:53:58.7556960Z * [new branch] gh/ezyang/3191/base -> origin/gh/ezyang/3191/base 2025-12-04T08:53:58.7557136Z * [new branch] gh/ezyang/3191/head -> origin/gh/ezyang/3191/head 2025-12-04T08:53:58.7557312Z * [new branch] gh/ezyang/3191/orig -> origin/gh/ezyang/3191/orig 2025-12-04T08:53:58.7557489Z * [new branch] gh/ezyang/3192/base -> origin/gh/ezyang/3192/base 2025-12-04T08:53:58.7557668Z * [new branch] gh/ezyang/3192/head -> origin/gh/ezyang/3192/head 2025-12-04T08:53:58.7557843Z * [new branch] gh/ezyang/3192/orig -> origin/gh/ezyang/3192/orig 2025-12-04T08:53:58.7558023Z * [new branch] gh/ezyang/3193/base -> origin/gh/ezyang/3193/base 2025-12-04T08:53:58.7558202Z * [new branch] gh/ezyang/3193/head -> origin/gh/ezyang/3193/head 2025-12-04T08:53:58.7558380Z * [new branch] gh/ezyang/3193/orig -> origin/gh/ezyang/3193/orig 2025-12-04T08:53:58.7558555Z * [new branch] gh/ezyang/3194/base -> origin/gh/ezyang/3194/base 2025-12-04T08:53:58.7558731Z * [new branch] gh/ezyang/3194/head -> origin/gh/ezyang/3194/head 2025-12-04T08:53:58.7558908Z * [new branch] gh/ezyang/3194/orig -> origin/gh/ezyang/3194/orig 2025-12-04T08:53:58.7559083Z * [new branch] gh/ezyang/3195/base -> origin/gh/ezyang/3195/base 2025-12-04T08:53:58.7559264Z * [new branch] gh/ezyang/3195/head -> origin/gh/ezyang/3195/head 2025-12-04T08:53:58.7559440Z * [new branch] gh/ezyang/3195/orig -> origin/gh/ezyang/3195/orig 2025-12-04T08:53:58.7559616Z * [new branch] gh/ezyang/3196/base -> origin/gh/ezyang/3196/base 2025-12-04T08:53:58.7559796Z * [new branch] gh/ezyang/3196/head -> origin/gh/ezyang/3196/head 2025-12-04T08:53:58.7559974Z * [new branch] gh/ezyang/3196/orig -> origin/gh/ezyang/3196/orig 2025-12-04T08:53:58.7560151Z * [new branch] gh/ezyang/3197/base -> origin/gh/ezyang/3197/base 2025-12-04T08:53:58.7560329Z * [new branch] gh/ezyang/3197/head -> origin/gh/ezyang/3197/head 2025-12-04T08:53:58.7560508Z * [new branch] gh/ezyang/3197/orig -> origin/gh/ezyang/3197/orig 2025-12-04T08:53:58.7560684Z * [new branch] gh/ezyang/3198/base -> origin/gh/ezyang/3198/base 2025-12-04T08:53:58.7560907Z * [new branch] gh/ezyang/3198/head -> origin/gh/ezyang/3198/head 2025-12-04T08:53:58.7561087Z * [new branch] gh/ezyang/3198/orig -> origin/gh/ezyang/3198/orig 2025-12-04T08:53:58.7561261Z * [new branch] gh/ezyang/3199/base -> origin/gh/ezyang/3199/base 2025-12-04T08:53:58.7561442Z * [new branch] gh/ezyang/3199/head -> origin/gh/ezyang/3199/head 2025-12-04T08:53:58.7561621Z * [new branch] gh/ezyang/3199/orig -> origin/gh/ezyang/3199/orig 2025-12-04T08:53:58.7561796Z * [new branch] gh/ezyang/3200/base -> origin/gh/ezyang/3200/base 2025-12-04T08:53:58.7561972Z * [new branch] gh/ezyang/3200/head -> origin/gh/ezyang/3200/head 2025-12-04T08:53:58.7562147Z * [new branch] gh/ezyang/3200/orig -> origin/gh/ezyang/3200/orig 2025-12-04T08:53:58.7562325Z * [new branch] gh/ezyang/3201/base -> origin/gh/ezyang/3201/base 2025-12-04T08:53:58.7562503Z * [new branch] gh/ezyang/3201/head -> origin/gh/ezyang/3201/head 2025-12-04T08:53:58.7562676Z * [new branch] gh/ezyang/3201/orig -> origin/gh/ezyang/3201/orig 2025-12-04T08:53:58.7562856Z * [new branch] gh/ezyang/3202/base -> origin/gh/ezyang/3202/base 2025-12-04T08:53:58.7563092Z * [new branch] gh/ezyang/3202/head -> origin/gh/ezyang/3202/head 2025-12-04T08:53:58.7563270Z * [new branch] gh/ezyang/3202/orig -> origin/gh/ezyang/3202/orig 2025-12-04T08:53:58.7563448Z * [new branch] gh/ezyang/3203/base -> origin/gh/ezyang/3203/base 2025-12-04T08:53:58.7563625Z * [new branch] gh/ezyang/3203/head -> origin/gh/ezyang/3203/head 2025-12-04T08:53:58.7563801Z * [new branch] gh/ezyang/3203/orig -> origin/gh/ezyang/3203/orig 2025-12-04T08:53:58.7563978Z * [new branch] gh/ezyang/3204/base -> origin/gh/ezyang/3204/base 2025-12-04T08:53:58.7564156Z * [new branch] gh/ezyang/3204/head -> origin/gh/ezyang/3204/head 2025-12-04T08:53:58.7564331Z * [new branch] gh/ezyang/3204/orig -> origin/gh/ezyang/3204/orig 2025-12-04T08:53:58.7564507Z * [new branch] gh/ezyang/3205/base -> origin/gh/ezyang/3205/base 2025-12-04T08:53:58.7564689Z * [new branch] gh/ezyang/3205/head -> origin/gh/ezyang/3205/head 2025-12-04T08:53:58.7564865Z * [new branch] gh/ezyang/3205/orig -> origin/gh/ezyang/3205/orig 2025-12-04T08:53:58.7565042Z * [new branch] gh/ezyang/3206/base -> origin/gh/ezyang/3206/base 2025-12-04T08:53:58.7565221Z * [new branch] gh/ezyang/3206/head -> origin/gh/ezyang/3206/head 2025-12-04T08:53:58.7565399Z * [new branch] gh/ezyang/3206/orig -> origin/gh/ezyang/3206/orig 2025-12-04T08:53:58.7565576Z * [new branch] gh/ezyang/3207/base -> origin/gh/ezyang/3207/base 2025-12-04T08:53:58.7565753Z * [new branch] gh/ezyang/3207/head -> origin/gh/ezyang/3207/head 2025-12-04T08:53:58.7565929Z * [new branch] gh/ezyang/3207/orig -> origin/gh/ezyang/3207/orig 2025-12-04T08:53:58.7566108Z * [new branch] gh/ezyang/3208/base -> origin/gh/ezyang/3208/base 2025-12-04T08:53:58.7566284Z * [new branch] gh/ezyang/3208/head -> origin/gh/ezyang/3208/head 2025-12-04T08:53:58.7566461Z * [new branch] gh/ezyang/3208/orig -> origin/gh/ezyang/3208/orig 2025-12-04T08:53:58.7566637Z * [new branch] gh/ezyang/3209/base -> origin/gh/ezyang/3209/base 2025-12-04T08:53:58.7566851Z * [new branch] gh/ezyang/3209/head -> origin/gh/ezyang/3209/head 2025-12-04T08:53:58.7567030Z * [new branch] gh/ezyang/3209/orig -> origin/gh/ezyang/3209/orig 2025-12-04T08:53:58.7567214Z * [new branch] gh/fadara01/3/base -> origin/gh/fadara01/3/base 2025-12-04T08:53:58.7567436Z * [new branch] gh/fadara01/3/head -> origin/gh/fadara01/3/head 2025-12-04T08:53:58.7567612Z * [new branch] gh/fadara01/3/orig -> origin/gh/fadara01/3/orig 2025-12-04T08:53:58.7567790Z * [new branch] gh/fadara01/5/base -> origin/gh/fadara01/5/base 2025-12-04T08:53:58.7567968Z * [new branch] gh/fadara01/5/head -> origin/gh/fadara01/5/head 2025-12-04T08:53:58.7568149Z * [new branch] gh/fadara01/5/orig -> origin/gh/fadara01/5/orig 2025-12-04T08:53:58.7568327Z * [new branch] gh/fadara01/6/base -> origin/gh/fadara01/6/base 2025-12-04T08:53:58.7568503Z * [new branch] gh/fadara01/6/head -> origin/gh/fadara01/6/head 2025-12-04T08:53:58.7568680Z * [new branch] gh/fadara01/6/orig -> origin/gh/fadara01/6/orig 2025-12-04T08:53:58.7568858Z * [new branch] gh/fadara01/7/base -> origin/gh/fadara01/7/base 2025-12-04T08:53:58.7569033Z * [new branch] gh/fadara01/7/head -> origin/gh/fadara01/7/head 2025-12-04T08:53:58.7569210Z * [new branch] gh/fadara01/7/orig -> origin/gh/fadara01/7/orig 2025-12-04T08:53:58.7569388Z * [new branch] gh/fadara01/8/base -> origin/gh/fadara01/8/base 2025-12-04T08:53:58.7569610Z * [new branch] gh/fadara01/8/head -> origin/gh/fadara01/8/head 2025-12-04T08:53:58.7569787Z * [new branch] gh/fadara01/8/orig -> origin/gh/fadara01/8/orig 2025-12-04T08:53:58.7569961Z * [new branch] gh/fadara01/9/base -> origin/gh/fadara01/9/base 2025-12-04T08:53:58.7570137Z * [new branch] gh/fadara01/9/head -> origin/gh/fadara01/9/head 2025-12-04T08:53:58.7570318Z * [new branch] gh/fadara01/9/orig -> origin/gh/fadara01/9/orig 2025-12-04T08:53:58.7570493Z * [new branch] gh/fduwjj/182/base -> origin/gh/fduwjj/182/base 2025-12-04T08:53:58.7570674Z * [new branch] gh/fduwjj/182/head -> origin/gh/fduwjj/182/head 2025-12-04T08:53:58.7570852Z * [new branch] gh/fduwjj/182/orig -> origin/gh/fduwjj/182/orig 2025-12-04T08:53:58.7571024Z * [new branch] gh/fduwjj/211/base -> origin/gh/fduwjj/211/base 2025-12-04T08:53:58.7571208Z * [new branch] gh/fduwjj/211/head -> origin/gh/fduwjj/211/head 2025-12-04T08:53:58.7571383Z * [new branch] gh/fduwjj/211/orig -> origin/gh/fduwjj/211/orig 2025-12-04T08:53:58.7571557Z * [new branch] gh/fduwjj/212/base -> origin/gh/fduwjj/212/base 2025-12-04T08:53:58.7571737Z * [new branch] gh/fduwjj/212/head -> origin/gh/fduwjj/212/head 2025-12-04T08:53:58.7571915Z * [new branch] gh/fduwjj/212/orig -> origin/gh/fduwjj/212/orig 2025-12-04T08:53:58.7572090Z * [new branch] gh/fduwjj/213/base -> origin/gh/fduwjj/213/base 2025-12-04T08:53:58.7572270Z * [new branch] gh/fduwjj/213/head -> origin/gh/fduwjj/213/head 2025-12-04T08:53:58.7572448Z * [new branch] gh/fduwjj/213/orig -> origin/gh/fduwjj/213/orig 2025-12-04T08:53:58.7572625Z * [new branch] gh/fduwjj/226/base -> origin/gh/fduwjj/226/base 2025-12-04T08:53:58.7572803Z * [new branch] gh/fduwjj/226/head -> origin/gh/fduwjj/226/head 2025-12-04T08:53:58.7572978Z * [new branch] gh/fduwjj/226/orig -> origin/gh/fduwjj/226/orig 2025-12-04T08:53:58.7573158Z * [new branch] gh/fduwjj/229/base -> origin/gh/fduwjj/229/base 2025-12-04T08:53:58.7573339Z * [new branch] gh/fduwjj/229/head -> origin/gh/fduwjj/229/head 2025-12-04T08:53:58.7573519Z * [new branch] gh/fduwjj/229/orig -> origin/gh/fduwjj/229/orig 2025-12-04T08:53:58.7573695Z * [new branch] gh/fduwjj/233/base -> origin/gh/fduwjj/233/base 2025-12-04T08:53:58.7573902Z * [new branch] gh/fduwjj/233/head -> origin/gh/fduwjj/233/head 2025-12-04T08:53:58.7574079Z * [new branch] gh/fduwjj/233/orig -> origin/gh/fduwjj/233/orig 2025-12-04T08:53:58.7574256Z * [new branch] gh/fduwjj/234/base -> origin/gh/fduwjj/234/base 2025-12-04T08:53:58.7574433Z * [new branch] gh/fduwjj/234/head -> origin/gh/fduwjj/234/head 2025-12-04T08:53:58.7574614Z * [new branch] gh/fduwjj/234/orig -> origin/gh/fduwjj/234/orig 2025-12-04T08:53:58.7574789Z * [new branch] gh/fduwjj/235/base -> origin/gh/fduwjj/235/base 2025-12-04T08:53:58.7574966Z * [new branch] gh/fduwjj/235/head -> origin/gh/fduwjj/235/head 2025-12-04T08:53:58.7575147Z * [new branch] gh/fduwjj/235/orig -> origin/gh/fduwjj/235/orig 2025-12-04T08:53:58.7575322Z * [new branch] gh/fduwjj/236/base -> origin/gh/fduwjj/236/base 2025-12-04T08:53:58.7575502Z * [new branch] gh/fduwjj/236/head -> origin/gh/fduwjj/236/head 2025-12-04T08:53:58.7575678Z * [new branch] gh/fduwjj/236/orig -> origin/gh/fduwjj/236/orig 2025-12-04T08:53:58.7575856Z * [new branch] gh/fduwjj/237/base -> origin/gh/fduwjj/237/base 2025-12-04T08:53:58.7576035Z * [new branch] gh/fduwjj/237/head -> origin/gh/fduwjj/237/head 2025-12-04T08:53:58.7576238Z * [new branch] gh/fduwjj/237/orig -> origin/gh/fduwjj/237/orig 2025-12-04T08:53:58.7576415Z * [new branch] gh/fduwjj/238/base -> origin/gh/fduwjj/238/base 2025-12-04T08:53:58.7576596Z * [new branch] gh/fduwjj/238/head -> origin/gh/fduwjj/238/head 2025-12-04T08:53:58.7576815Z * [new branch] gh/fduwjj/238/orig -> origin/gh/fduwjj/238/orig 2025-12-04T08:53:58.7576994Z * [new branch] gh/fduwjj/239/base -> origin/gh/fduwjj/239/base 2025-12-04T08:53:58.7577174Z * [new branch] gh/fduwjj/239/head -> origin/gh/fduwjj/239/head 2025-12-04T08:53:58.7577351Z * [new branch] gh/fduwjj/239/orig -> origin/gh/fduwjj/239/orig 2025-12-04T08:53:58.7577535Z * [new branch] gh/fegin/332/base -> origin/gh/fegin/332/base 2025-12-04T08:53:58.7577720Z * [new branch] gh/fegin/332/head -> origin/gh/fegin/332/head 2025-12-04T08:53:58.7577900Z * [new branch] gh/fegin/332/orig -> origin/gh/fegin/332/orig 2025-12-04T08:53:58.7578082Z * [new branch] gh/fegin/333/base -> origin/gh/fegin/333/base 2025-12-04T08:53:58.7578259Z * [new branch] gh/fegin/333/head -> origin/gh/fegin/333/head 2025-12-04T08:53:58.7578432Z * [new branch] gh/fegin/333/orig -> origin/gh/fegin/333/orig 2025-12-04T08:53:58.7578605Z * [new branch] gh/fegin/334/base -> origin/gh/fegin/334/base 2025-12-04T08:53:58.7578783Z * [new branch] gh/fegin/334/head -> origin/gh/fegin/334/head 2025-12-04T08:53:58.7578957Z * [new branch] gh/fegin/334/orig -> origin/gh/fegin/334/orig 2025-12-04T08:53:58.7579138Z * [new branch] gh/fegin/335/base -> origin/gh/fegin/335/base 2025-12-04T08:53:58.7579314Z * [new branch] gh/fegin/335/head -> origin/gh/fegin/335/head 2025-12-04T08:53:58.7579496Z * [new branch] gh/fegin/335/orig -> origin/gh/fegin/335/orig 2025-12-04T08:53:58.7579678Z * [new branch] gh/fffrog/160/base -> origin/gh/fffrog/160/base 2025-12-04T08:53:58.7579856Z * [new branch] gh/fffrog/160/head -> origin/gh/fffrog/160/head 2025-12-04T08:53:58.7580038Z * [new branch] gh/fffrog/177/base -> origin/gh/fffrog/177/base 2025-12-04T08:53:58.7580223Z * [new branch] gh/fffrog/177/head -> origin/gh/fffrog/177/head 2025-12-04T08:53:58.7580441Z * [new branch] gh/fffrog/177/orig -> origin/gh/fffrog/177/orig 2025-12-04T08:53:58.7580620Z * [new branch] gh/fffrog/178/base -> origin/gh/fffrog/178/base 2025-12-04T08:53:58.7580801Z * [new branch] gh/fffrog/178/head -> origin/gh/fffrog/178/head 2025-12-04T08:53:58.7580980Z * [new branch] gh/fffrog/178/orig -> origin/gh/fffrog/178/orig 2025-12-04T08:53:58.7581162Z * [new branch] gh/fffrog/181/base -> origin/gh/fffrog/181/base 2025-12-04T08:53:58.7581343Z * [new branch] gh/fffrog/181/head -> origin/gh/fffrog/181/head 2025-12-04T08:53:58.7581520Z * [new branch] gh/fffrog/181/orig -> origin/gh/fffrog/181/orig 2025-12-04T08:53:58.7581702Z * [new branch] gh/fffrog/183/base -> origin/gh/fffrog/183/base 2025-12-04T08:53:58.7581884Z * [new branch] gh/fffrog/183/head -> origin/gh/fffrog/183/head 2025-12-04T08:53:58.7582068Z * [new branch] gh/fffrog/183/orig -> origin/gh/fffrog/183/orig 2025-12-04T08:53:58.7582249Z * [new branch] gh/fxdawnn/10/base -> origin/gh/fxdawnn/10/base 2025-12-04T08:53:58.7582431Z * [new branch] gh/fxdawnn/10/head -> origin/gh/fxdawnn/10/head 2025-12-04T08:53:58.7582610Z * [new branch] gh/fxdawnn/10/orig -> origin/gh/fxdawnn/10/orig 2025-12-04T08:53:58.7582841Z * [new branch] gh/fxdawnn/11/base -> origin/gh/fxdawnn/11/base 2025-12-04T08:53:58.7583023Z * [new branch] gh/fxdawnn/11/head -> origin/gh/fxdawnn/11/head 2025-12-04T08:53:58.7583205Z * [new branch] gh/fxdawnn/11/orig -> origin/gh/fxdawnn/11/orig 2025-12-04T08:53:58.7583388Z * [new branch] gh/fxdawnn/12/base -> origin/gh/fxdawnn/12/base 2025-12-04T08:53:58.7583568Z * [new branch] gh/fxdawnn/12/head -> origin/gh/fxdawnn/12/head 2025-12-04T08:53:58.7583752Z * [new branch] gh/fxdawnn/12/orig -> origin/gh/fxdawnn/12/orig 2025-12-04T08:53:58.7583939Z * [new branch] gh/fxdawnn/13/base -> origin/gh/fxdawnn/13/base 2025-12-04T08:53:58.7584118Z * [new branch] gh/fxdawnn/13/head -> origin/gh/fxdawnn/13/head 2025-12-04T08:53:58.7584304Z * [new branch] gh/fxdawnn/13/orig -> origin/gh/fxdawnn/13/orig 2025-12-04T08:53:58.7584492Z * [new branch] gh/fxdawnn/14/base -> origin/gh/fxdawnn/14/base 2025-12-04T08:53:58.7584686Z * [new branch] gh/fxdawnn/14/head -> origin/gh/fxdawnn/14/head 2025-12-04T08:53:58.7584868Z * [new branch] gh/fxdawnn/14/orig -> origin/gh/fxdawnn/14/orig 2025-12-04T08:53:58.7585052Z * [new branch] gh/fxdawnn/15/base -> origin/gh/fxdawnn/15/base 2025-12-04T08:53:58.7585232Z * [new branch] gh/fxdawnn/15/head -> origin/gh/fxdawnn/15/head 2025-12-04T08:53:58.7585416Z * [new branch] gh/fxdawnn/15/orig -> origin/gh/fxdawnn/15/orig 2025-12-04T08:53:58.7585603Z * [new branch] gh/fxdawnn/6/base -> origin/gh/fxdawnn/6/base 2025-12-04T08:53:58.7585782Z * [new branch] gh/fxdawnn/6/head -> origin/gh/fxdawnn/6/head 2025-12-04T08:53:58.7585964Z * [new branch] gh/fxdawnn/6/orig -> origin/gh/fxdawnn/6/orig 2025-12-04T08:53:58.7586141Z * [new branch] gh/fxdawnn/7/base -> origin/gh/fxdawnn/7/base 2025-12-04T08:53:58.7586320Z * [new branch] gh/fxdawnn/7/head -> origin/gh/fxdawnn/7/head 2025-12-04T08:53:58.7586501Z * [new branch] gh/fxdawnn/7/orig -> origin/gh/fxdawnn/7/orig 2025-12-04T08:53:58.7586678Z * [new branch] gh/fxdawnn/9/base -> origin/gh/fxdawnn/9/base 2025-12-04T08:53:58.7586905Z * [new branch] gh/fxdawnn/9/head -> origin/gh/fxdawnn/9/head 2025-12-04T08:53:58.7587084Z * [new branch] gh/fxdawnn/9/orig -> origin/gh/fxdawnn/9/orig 2025-12-04T08:53:58.7587318Z * [new branch] gh/galv/1/base -> origin/gh/galv/1/base 2025-12-04T08:53:58.7587492Z * [new branch] gh/galv/1/head -> origin/gh/galv/1/head 2025-12-04T08:53:58.7587663Z * [new branch] gh/galv/1/orig -> origin/gh/galv/1/orig 2025-12-04T08:53:58.7587831Z * [new branch] gh/galv/2/base -> origin/gh/galv/2/base 2025-12-04T08:53:58.7588002Z * [new branch] gh/galv/2/head -> origin/gh/galv/2/head 2025-12-04T08:53:58.7588170Z * [new branch] gh/galv/2/orig -> origin/gh/galv/2/orig 2025-12-04T08:53:58.7588335Z * [new branch] gh/galv/3/base -> origin/gh/galv/3/base 2025-12-04T08:53:58.7588502Z * [new branch] gh/galv/3/head -> origin/gh/galv/3/head 2025-12-04T08:53:58.7588669Z * [new branch] gh/galv/3/orig -> origin/gh/galv/3/orig 2025-12-04T08:53:58.7588851Z * [new branch] gh/guangyey/134/base -> origin/gh/guangyey/134/base 2025-12-04T08:53:58.7589040Z * [new branch] gh/guangyey/134/head -> origin/gh/guangyey/134/head 2025-12-04T08:53:58.7589225Z * [new branch] gh/guangyey/134/orig -> origin/gh/guangyey/134/orig 2025-12-04T08:53:58.7589410Z * [new branch] gh/guangyey/163/base -> origin/gh/guangyey/163/base 2025-12-04T08:53:58.7589632Z * [new branch] gh/guangyey/163/head -> origin/gh/guangyey/163/head 2025-12-04T08:53:58.7589816Z * [new branch] gh/guangyey/163/orig -> origin/gh/guangyey/163/orig 2025-12-04T08:53:58.7589998Z * [new branch] gh/guangyey/168/base -> origin/gh/guangyey/168/base 2025-12-04T08:53:58.7590181Z * [new branch] gh/guangyey/168/head -> origin/gh/guangyey/168/head 2025-12-04T08:53:58.7590362Z * [new branch] gh/guangyey/168/orig -> origin/gh/guangyey/168/orig 2025-12-04T08:53:58.7590547Z * [new branch] gh/guangyey/169/base -> origin/gh/guangyey/169/base 2025-12-04T08:53:58.7590730Z * [new branch] gh/guangyey/169/head -> origin/gh/guangyey/169/head 2025-12-04T08:53:58.7590912Z * [new branch] gh/guangyey/169/orig -> origin/gh/guangyey/169/orig 2025-12-04T08:53:58.7591097Z * [new branch] gh/guangyey/170/base -> origin/gh/guangyey/170/base 2025-12-04T08:53:58.7591281Z * [new branch] gh/guangyey/170/head -> origin/gh/guangyey/170/head 2025-12-04T08:53:58.7591462Z * [new branch] gh/guangyey/170/orig -> origin/gh/guangyey/170/orig 2025-12-04T08:53:58.7591649Z * [new branch] gh/guangyey/171/base -> origin/gh/guangyey/171/base 2025-12-04T08:53:58.7591835Z * [new branch] gh/guangyey/171/head -> origin/gh/guangyey/171/head 2025-12-04T08:53:58.7592017Z * [new branch] gh/guangyey/171/orig -> origin/gh/guangyey/171/orig 2025-12-04T08:53:58.7592202Z * [new branch] gh/guangyey/178/base -> origin/gh/guangyey/178/base 2025-12-04T08:53:58.7592384Z * [new branch] gh/guangyey/178/head -> origin/gh/guangyey/178/head 2025-12-04T08:53:58.7592565Z * [new branch] gh/guangyey/178/orig -> origin/gh/guangyey/178/orig 2025-12-04T08:53:58.7592751Z * [new branch] gh/guangyey/182/base -> origin/gh/guangyey/182/base 2025-12-04T08:53:58.7592935Z * [new branch] gh/guangyey/182/head -> origin/gh/guangyey/182/head 2025-12-04T08:53:58.7593116Z * [new branch] gh/guangyey/182/orig -> origin/gh/guangyey/182/orig 2025-12-04T08:53:58.7593299Z * [new branch] gh/guangyey/183/base -> origin/gh/guangyey/183/base 2025-12-04T08:53:58.7593481Z * [new branch] gh/guangyey/183/head -> origin/gh/guangyey/183/head 2025-12-04T08:53:58.7593662Z * [new branch] gh/guangyey/183/orig -> origin/gh/guangyey/183/orig 2025-12-04T08:53:58.7593905Z * [new branch] gh/guangyey/185/base -> origin/gh/guangyey/185/base 2025-12-04T08:53:58.7594085Z * [new branch] gh/guangyey/185/head -> origin/gh/guangyey/185/head 2025-12-04T08:53:58.7594269Z * [new branch] gh/guangyey/185/orig -> origin/gh/guangyey/185/orig 2025-12-04T08:53:58.7594454Z * [new branch] gh/guangyey/186/base -> origin/gh/guangyey/186/base 2025-12-04T08:53:58.7594634Z * [new branch] gh/guangyey/186/head -> origin/gh/guangyey/186/head 2025-12-04T08:53:58.7594816Z * [new branch] gh/guangyey/186/orig -> origin/gh/guangyey/186/orig 2025-12-04T08:53:58.7595000Z * [new branch] gh/guangyey/187/base -> origin/gh/guangyey/187/base 2025-12-04T08:53:58.7595182Z * [new branch] gh/guangyey/187/head -> origin/gh/guangyey/187/head 2025-12-04T08:53:58.7595363Z * [new branch] gh/guangyey/187/orig -> origin/gh/guangyey/187/orig 2025-12-04T08:53:58.7595547Z * [new branch] gh/guangyey/188/base -> origin/gh/guangyey/188/base 2025-12-04T08:53:58.7595730Z * [new branch] gh/guangyey/188/head -> origin/gh/guangyey/188/head 2025-12-04T08:53:58.7595912Z * [new branch] gh/guangyey/188/orig -> origin/gh/guangyey/188/orig 2025-12-04T08:53:58.7596127Z * [new branch] gh/guangyey/190/base -> origin/gh/guangyey/190/base 2025-12-04T08:53:58.7596307Z * [new branch] gh/guangyey/190/head -> origin/gh/guangyey/190/head 2025-12-04T08:53:58.7596491Z * [new branch] gh/guangyey/190/orig -> origin/gh/guangyey/190/orig 2025-12-04T08:53:58.7596673Z * [new branch] gh/guangyey/208/base -> origin/gh/guangyey/208/base 2025-12-04T08:53:58.7596894Z * [new branch] gh/guangyey/208/head -> origin/gh/guangyey/208/head 2025-12-04T08:53:58.7597080Z * [new branch] gh/guangyey/208/orig -> origin/gh/guangyey/208/orig 2025-12-04T08:53:58.7597267Z * [new branch] gh/guangyey/228/base -> origin/gh/guangyey/228/base 2025-12-04T08:53:58.7597446Z * [new branch] gh/guangyey/228/head -> origin/gh/guangyey/228/head 2025-12-04T08:53:58.7597630Z * [new branch] gh/guangyey/228/orig -> origin/gh/guangyey/228/orig 2025-12-04T08:53:58.7597819Z * [new branch] gh/guangyey/230/base -> origin/gh/guangyey/230/base 2025-12-04T08:53:58.7598001Z * [new branch] gh/guangyey/230/head -> origin/gh/guangyey/230/head 2025-12-04T08:53:58.7598182Z * [new branch] gh/guangyey/230/orig -> origin/gh/guangyey/230/orig 2025-12-04T08:53:58.7598363Z * [new branch] gh/guangyey/231/base -> origin/gh/guangyey/231/base 2025-12-04T08:53:58.7598546Z * [new branch] gh/guangyey/231/head -> origin/gh/guangyey/231/head 2025-12-04T08:53:58.7598729Z * [new branch] gh/guangyey/231/orig -> origin/gh/guangyey/231/orig 2025-12-04T08:53:58.7598911Z * [new branch] gh/guangyey/232/base -> origin/gh/guangyey/232/base 2025-12-04T08:53:58.7599096Z * [new branch] gh/guangyey/232/head -> origin/gh/guangyey/232/head 2025-12-04T08:53:58.7599278Z * [new branch] gh/guangyey/232/orig -> origin/gh/guangyey/232/orig 2025-12-04T08:53:58.7599464Z * [new branch] gh/guangyey/233/base -> origin/gh/guangyey/233/base 2025-12-04T08:53:58.7599652Z * [new branch] gh/guangyey/233/head -> origin/gh/guangyey/233/head 2025-12-04T08:53:58.7599836Z * [new branch] gh/guangyey/233/orig -> origin/gh/guangyey/233/orig 2025-12-04T08:53:58.7600016Z * [new branch] gh/guangyey/234/base -> origin/gh/guangyey/234/base 2025-12-04T08:53:58.7600200Z * [new branch] gh/guangyey/234/head -> origin/gh/guangyey/234/head 2025-12-04T08:53:58.7600428Z * [new branch] gh/guangyey/234/orig -> origin/gh/guangyey/234/orig 2025-12-04T08:53:58.7600608Z * [new branch] gh/guangyey/235/base -> origin/gh/guangyey/235/base 2025-12-04T08:53:58.7600790Z * [new branch] gh/guangyey/235/head -> origin/gh/guangyey/235/head 2025-12-04T08:53:58.7600973Z * [new branch] gh/guangyey/235/orig -> origin/gh/guangyey/235/orig 2025-12-04T08:53:58.7601156Z * [new branch] gh/guangyey/236/base -> origin/gh/guangyey/236/base 2025-12-04T08:53:58.7601343Z * [new branch] gh/guangyey/236/head -> origin/gh/guangyey/236/head 2025-12-04T08:53:58.7601526Z * [new branch] gh/guangyey/236/orig -> origin/gh/guangyey/236/orig 2025-12-04T08:53:58.7601707Z * [new branch] gh/guangyey/237/base -> origin/gh/guangyey/237/base 2025-12-04T08:53:58.7601888Z * [new branch] gh/guangyey/237/head -> origin/gh/guangyey/237/head 2025-12-04T08:53:58.7602073Z * [new branch] gh/guangyey/237/orig -> origin/gh/guangyey/237/orig 2025-12-04T08:53:58.7602258Z * [new branch] gh/guangyey/238/base -> origin/gh/guangyey/238/base 2025-12-04T08:53:58.7602443Z * [new branch] gh/guangyey/238/head -> origin/gh/guangyey/238/head 2025-12-04T08:53:58.7602625Z * [new branch] gh/guangyey/239/base -> origin/gh/guangyey/239/base 2025-12-04T08:53:58.7602865Z * [new branch] gh/guangyey/239/head -> origin/gh/guangyey/239/head 2025-12-04T08:53:58.7603056Z * [new branch] gh/guangyey/239/orig -> origin/gh/guangyey/239/orig 2025-12-04T08:53:58.7603237Z * [new branch] gh/guangyey/240/base -> origin/gh/guangyey/240/base 2025-12-04T08:53:58.7603419Z * [new branch] gh/guangyey/240/head -> origin/gh/guangyey/240/head 2025-12-04T08:53:58.7603604Z * [new branch] gh/guangyey/240/orig -> origin/gh/guangyey/240/orig 2025-12-04T08:53:58.7603787Z * [new branch] gh/guangyey/241/base -> origin/gh/guangyey/241/base 2025-12-04T08:53:58.7603971Z * [new branch] gh/guangyey/241/head -> origin/gh/guangyey/241/head 2025-12-04T08:53:58.7604157Z * [new branch] gh/guangyey/241/orig -> origin/gh/guangyey/241/orig 2025-12-04T08:53:58.7604336Z * [new branch] gh/guangyey/242/base -> origin/gh/guangyey/242/base 2025-12-04T08:53:58.7604522Z * [new branch] gh/guangyey/242/head -> origin/gh/guangyey/242/head 2025-12-04T08:53:58.7604708Z * [new branch] gh/guangyey/242/orig -> origin/gh/guangyey/242/orig 2025-12-04T08:53:58.7604891Z * [new branch] gh/guangyey/243/base -> origin/gh/guangyey/243/base 2025-12-04T08:53:58.7605073Z * [new branch] gh/guangyey/243/head -> origin/gh/guangyey/243/head 2025-12-04T08:53:58.7605257Z * [new branch] gh/guangyey/243/orig -> origin/gh/guangyey/243/orig 2025-12-04T08:53:58.7605439Z * [new branch] gh/guangyey/244/base -> origin/gh/guangyey/244/base 2025-12-04T08:53:58.7605625Z * [new branch] gh/guangyey/244/head -> origin/gh/guangyey/244/head 2025-12-04T08:53:58.7605808Z * [new branch] gh/guangyey/244/orig -> origin/gh/guangyey/244/orig 2025-12-04T08:53:58.7605988Z * [new branch] gh/guangyey/245/base -> origin/gh/guangyey/245/base 2025-12-04T08:53:58.7606172Z * [new branch] gh/guangyey/245/head -> origin/gh/guangyey/245/head 2025-12-04T08:53:58.7606353Z * [new branch] gh/guangyey/245/orig -> origin/gh/guangyey/245/orig 2025-12-04T08:53:58.7606537Z * [new branch] gh/guangyey/246/base -> origin/gh/guangyey/246/base 2025-12-04T08:53:58.7606725Z * [new branch] gh/guangyey/246/head -> origin/gh/guangyey/246/head 2025-12-04T08:53:58.7606956Z * [new branch] gh/guangyey/246/orig -> origin/gh/guangyey/246/orig 2025-12-04T08:53:58.7607182Z * [new branch] gh/guangyey/247/base -> origin/gh/guangyey/247/base 2025-12-04T08:53:58.7607365Z * [new branch] gh/guangyey/247/head -> origin/gh/guangyey/247/head 2025-12-04T08:53:58.7607550Z * [new branch] gh/guangyey/247/orig -> origin/gh/guangyey/247/orig 2025-12-04T08:53:58.7607749Z * [new branch] gh/guangyey/248/base -> origin/gh/guangyey/248/base 2025-12-04T08:53:58.7607932Z * [new branch] gh/guangyey/248/head -> origin/gh/guangyey/248/head 2025-12-04T08:53:58.7608005Z * [new branch] gh/guangyey/248/orig -> origin/gh/guangyey/248/orig 2025-12-04T08:53:58.7608076Z * [new branch] gh/guangyey/249/base -> origin/gh/guangyey/249/base 2025-12-04T08:53:58.7608150Z * [new branch] gh/guangyey/249/head -> origin/gh/guangyey/249/head 2025-12-04T08:53:58.7608221Z * [new branch] gh/guangyey/249/orig -> origin/gh/guangyey/249/orig 2025-12-04T08:53:58.7608295Z * [new branch] gh/guangyey/250/base -> origin/gh/guangyey/250/base 2025-12-04T08:53:58.7608372Z * [new branch] gh/guangyey/250/head -> origin/gh/guangyey/250/head 2025-12-04T08:53:58.7608444Z * [new branch] gh/guangyey/250/orig -> origin/gh/guangyey/250/orig 2025-12-04T08:53:58.7608563Z * [new branch] gh/guangyey/251/base -> origin/gh/guangyey/251/base 2025-12-04T08:53:58.7608639Z * [new branch] gh/guangyey/251/head -> origin/gh/guangyey/251/head 2025-12-04T08:53:58.7608712Z * [new branch] gh/guangyey/251/orig -> origin/gh/guangyey/251/orig 2025-12-04T08:53:58.7608783Z * [new branch] gh/guangyey/252/base -> origin/gh/guangyey/252/base 2025-12-04T08:53:58.7608857Z * [new branch] gh/guangyey/252/head -> origin/gh/guangyey/252/head 2025-12-04T08:53:58.7608931Z * [new branch] gh/guangyey/252/orig -> origin/gh/guangyey/252/orig 2025-12-04T08:53:58.7609007Z * [new branch] gh/guangyey/253/base -> origin/gh/guangyey/253/base 2025-12-04T08:53:58.7609077Z * [new branch] gh/guangyey/253/head -> origin/gh/guangyey/253/head 2025-12-04T08:53:58.7609149Z * [new branch] gh/guangyey/253/orig -> origin/gh/guangyey/253/orig 2025-12-04T08:53:58.7609225Z * [new branch] gh/guangyey/254/base -> origin/gh/guangyey/254/base 2025-12-04T08:53:58.7609297Z * [new branch] gh/guangyey/254/head -> origin/gh/guangyey/254/head 2025-12-04T08:53:58.7609369Z * [new branch] gh/guangyey/254/orig -> origin/gh/guangyey/254/orig 2025-12-04T08:53:58.7609443Z * [new branch] gh/guangyey/255/base -> origin/gh/guangyey/255/base 2025-12-04T08:53:58.7609514Z * [new branch] gh/guangyey/255/head -> origin/gh/guangyey/255/head 2025-12-04T08:53:58.7609586Z * [new branch] gh/guangyey/255/orig -> origin/gh/guangyey/255/orig 2025-12-04T08:53:58.7609688Z * [new branch] gh/guilhermeleobas/107/base -> origin/gh/guilhermeleobas/107/base 2025-12-04T08:53:58.7609781Z * [new branch] gh/guilhermeleobas/107/head -> origin/gh/guilhermeleobas/107/head 2025-12-04T08:53:58.7609872Z * [new branch] gh/guilhermeleobas/107/orig -> origin/gh/guilhermeleobas/107/orig 2025-12-04T08:53:58.7609968Z * [new branch] gh/guilhermeleobas/108/base -> origin/gh/guilhermeleobas/108/base 2025-12-04T08:53:58.7610056Z * [new branch] gh/guilhermeleobas/108/head -> origin/gh/guilhermeleobas/108/head 2025-12-04T08:53:58.7610146Z * [new branch] gh/guilhermeleobas/108/orig -> origin/gh/guilhermeleobas/108/orig 2025-12-04T08:53:58.7610238Z * [new branch] gh/guilhermeleobas/150/base -> origin/gh/guilhermeleobas/150/base 2025-12-04T08:53:58.7610326Z * [new branch] gh/guilhermeleobas/150/head -> origin/gh/guilhermeleobas/150/head 2025-12-04T08:53:58.7610453Z * [new branch] gh/guilhermeleobas/150/orig -> origin/gh/guilhermeleobas/150/orig 2025-12-04T08:53:58.7610543Z * [new branch] gh/guilhermeleobas/168/base -> origin/gh/guilhermeleobas/168/base 2025-12-04T08:53:58.7610634Z * [new branch] gh/guilhermeleobas/168/head -> origin/gh/guilhermeleobas/168/head 2025-12-04T08:53:58.7610726Z * [new branch] gh/guilhermeleobas/168/orig -> origin/gh/guilhermeleobas/168/orig 2025-12-04T08:53:58.7610814Z * [new branch] gh/guilhermeleobas/169/base -> origin/gh/guilhermeleobas/169/base 2025-12-04T08:53:58.7610902Z * [new branch] gh/guilhermeleobas/169/head -> origin/gh/guilhermeleobas/169/head 2025-12-04T08:53:58.7610991Z * [new branch] gh/guilhermeleobas/169/orig -> origin/gh/guilhermeleobas/169/orig 2025-12-04T08:53:58.7611081Z * [new branch] gh/guilhermeleobas/170/base -> origin/gh/guilhermeleobas/170/base 2025-12-04T08:53:58.7611173Z * [new branch] gh/guilhermeleobas/170/head -> origin/gh/guilhermeleobas/170/head 2025-12-04T08:53:58.7611265Z * [new branch] gh/guilhermeleobas/170/orig -> origin/gh/guilhermeleobas/170/orig 2025-12-04T08:53:58.7611354Z * [new branch] gh/guilhermeleobas/171/base -> origin/gh/guilhermeleobas/171/base 2025-12-04T08:53:58.7611465Z * [new branch] gh/guilhermeleobas/171/head -> origin/gh/guilhermeleobas/171/head 2025-12-04T08:53:58.7611554Z * [new branch] gh/guilhermeleobas/171/orig -> origin/gh/guilhermeleobas/171/orig 2025-12-04T08:53:58.7611644Z * [new branch] gh/guilhermeleobas/173/base -> origin/gh/guilhermeleobas/173/base 2025-12-04T08:53:58.7611735Z * [new branch] gh/guilhermeleobas/173/head -> origin/gh/guilhermeleobas/173/head 2025-12-04T08:53:58.7611823Z * [new branch] gh/guilhermeleobas/173/orig -> origin/gh/guilhermeleobas/173/orig 2025-12-04T08:53:58.7611913Z * [new branch] gh/guilhermeleobas/193/base -> origin/gh/guilhermeleobas/193/base 2025-12-04T08:53:58.7612005Z * [new branch] gh/guilhermeleobas/193/head -> origin/gh/guilhermeleobas/193/head 2025-12-04T08:53:58.7612093Z * [new branch] gh/guilhermeleobas/193/orig -> origin/gh/guilhermeleobas/193/orig 2025-12-04T08:53:58.7612184Z * [new branch] gh/guilhermeleobas/204/base -> origin/gh/guilhermeleobas/204/base 2025-12-04T08:53:58.7612274Z * [new branch] gh/guilhermeleobas/204/head -> origin/gh/guilhermeleobas/204/head 2025-12-04T08:53:58.7612362Z * [new branch] gh/guilhermeleobas/204/orig -> origin/gh/guilhermeleobas/204/orig 2025-12-04T08:53:58.7612450Z * [new branch] gh/guilhermeleobas/211/base -> origin/gh/guilhermeleobas/211/base 2025-12-04T08:53:58.7612540Z * [new branch] gh/guilhermeleobas/211/head -> origin/gh/guilhermeleobas/211/head 2025-12-04T08:53:58.7612630Z * [new branch] gh/guilhermeleobas/211/orig -> origin/gh/guilhermeleobas/211/orig 2025-12-04T08:53:58.7612720Z * [new branch] gh/guilhermeleobas/226/base -> origin/gh/guilhermeleobas/226/base 2025-12-04T08:53:58.7612811Z * [new branch] gh/guilhermeleobas/226/head -> origin/gh/guilhermeleobas/226/head 2025-12-04T08:53:58.7612901Z * [new branch] gh/guilhermeleobas/226/orig -> origin/gh/guilhermeleobas/226/orig 2025-12-04T08:53:58.7612989Z * [new branch] gh/guilhermeleobas/236/base -> origin/gh/guilhermeleobas/236/base 2025-12-04T08:53:58.7613079Z * [new branch] gh/guilhermeleobas/236/head -> origin/gh/guilhermeleobas/236/head 2025-12-04T08:53:58.7613169Z * [new branch] gh/guilhermeleobas/236/orig -> origin/gh/guilhermeleobas/236/orig 2025-12-04T08:53:58.7613261Z * [new branch] gh/guilhermeleobas/247/base -> origin/gh/guilhermeleobas/247/base 2025-12-04T08:53:58.7613376Z * [new branch] gh/guilhermeleobas/247/head -> origin/gh/guilhermeleobas/247/head 2025-12-04T08:53:58.7613465Z * [new branch] gh/guilhermeleobas/247/orig -> origin/gh/guilhermeleobas/247/orig 2025-12-04T08:53:58.7613557Z * [new branch] gh/guilhermeleobas/248/base -> origin/gh/guilhermeleobas/248/base 2025-12-04T08:53:58.7613645Z * [new branch] gh/guilhermeleobas/248/head -> origin/gh/guilhermeleobas/248/head 2025-12-04T08:53:58.7613736Z * [new branch] gh/guilhermeleobas/248/orig -> origin/gh/guilhermeleobas/248/orig 2025-12-04T08:53:58.7613827Z * [new branch] gh/guilhermeleobas/250/base -> origin/gh/guilhermeleobas/250/base 2025-12-04T08:53:58.7613916Z * [new branch] gh/guilhermeleobas/250/head -> origin/gh/guilhermeleobas/250/head 2025-12-04T08:53:58.7614005Z * [new branch] gh/guilhermeleobas/250/orig -> origin/gh/guilhermeleobas/250/orig 2025-12-04T08:53:58.7614097Z * [new branch] gh/guilhermeleobas/253/base -> origin/gh/guilhermeleobas/253/base 2025-12-04T08:53:58.7614188Z * [new branch] gh/guilhermeleobas/253/head -> origin/gh/guilhermeleobas/253/head 2025-12-04T08:53:58.7614277Z * [new branch] gh/guilhermeleobas/253/orig -> origin/gh/guilhermeleobas/253/orig 2025-12-04T08:53:58.7614369Z * [new branch] gh/guilhermeleobas/254/base -> origin/gh/guilhermeleobas/254/base 2025-12-04T08:53:58.7614483Z * [new branch] gh/guilhermeleobas/254/head -> origin/gh/guilhermeleobas/254/head 2025-12-04T08:53:58.7614573Z * [new branch] gh/guilhermeleobas/254/orig -> origin/gh/guilhermeleobas/254/orig 2025-12-04T08:53:58.7614661Z * [new branch] gh/guilhermeleobas/255/base -> origin/gh/guilhermeleobas/255/base 2025-12-04T08:53:58.7614750Z * [new branch] gh/guilhermeleobas/255/head -> origin/gh/guilhermeleobas/255/head 2025-12-04T08:53:58.7614842Z * [new branch] gh/guilhermeleobas/255/orig -> origin/gh/guilhermeleobas/255/orig 2025-12-04T08:53:58.7614931Z * [new branch] gh/guilhermeleobas/256/base -> origin/gh/guilhermeleobas/256/base 2025-12-04T08:53:58.7615021Z * [new branch] gh/guilhermeleobas/256/head -> origin/gh/guilhermeleobas/256/head 2025-12-04T08:53:58.7615113Z * [new branch] gh/guilhermeleobas/256/orig -> origin/gh/guilhermeleobas/256/orig 2025-12-04T08:53:58.7615204Z * [new branch] gh/guilhermeleobas/257/base -> origin/gh/guilhermeleobas/257/base 2025-12-04T08:53:58.7615293Z * [new branch] gh/guilhermeleobas/257/head -> origin/gh/guilhermeleobas/257/head 2025-12-04T08:53:58.7615383Z * [new branch] gh/guilhermeleobas/257/orig -> origin/gh/guilhermeleobas/257/orig 2025-12-04T08:53:58.7615473Z * [new branch] gh/guilhermeleobas/258/base -> origin/gh/guilhermeleobas/258/base 2025-12-04T08:53:58.7615562Z * [new branch] gh/guilhermeleobas/258/head -> origin/gh/guilhermeleobas/258/head 2025-12-04T08:53:58.7615654Z * [new branch] gh/guilhermeleobas/258/orig -> origin/gh/guilhermeleobas/258/orig 2025-12-04T08:53:58.7615743Z * [new branch] gh/guilhermeleobas/259/base -> origin/gh/guilhermeleobas/259/base 2025-12-04T08:53:58.7615832Z * [new branch] gh/guilhermeleobas/259/head -> origin/gh/guilhermeleobas/259/head 2025-12-04T08:53:58.7615925Z * [new branch] gh/guilhermeleobas/259/orig -> origin/gh/guilhermeleobas/259/orig 2025-12-04T08:53:58.7616015Z * [new branch] gh/guilhermeleobas/260/base -> origin/gh/guilhermeleobas/260/base 2025-12-04T08:53:58.7616102Z * [new branch] gh/guilhermeleobas/260/head -> origin/gh/guilhermeleobas/260/head 2025-12-04T08:53:58.7616191Z * [new branch] gh/guilhermeleobas/260/orig -> origin/gh/guilhermeleobas/260/orig 2025-12-04T08:53:58.7616279Z * [new branch] gh/guilhermeleobas/261/base -> origin/gh/guilhermeleobas/261/base 2025-12-04T08:53:58.7616393Z * [new branch] gh/guilhermeleobas/261/head -> origin/gh/guilhermeleobas/261/head 2025-12-04T08:53:58.7616481Z * [new branch] gh/guilhermeleobas/261/orig -> origin/gh/guilhermeleobas/261/orig 2025-12-04T08:53:58.7616570Z * [new branch] gh/guilhermeleobas/262/base -> origin/gh/guilhermeleobas/262/base 2025-12-04T08:53:58.7616662Z * [new branch] gh/guilhermeleobas/262/head -> origin/gh/guilhermeleobas/262/head 2025-12-04T08:53:58.7616819Z * [new branch] gh/guilhermeleobas/262/orig -> origin/gh/guilhermeleobas/262/orig 2025-12-04T08:53:58.7616910Z * [new branch] gh/guilhermeleobas/263/base -> origin/gh/guilhermeleobas/263/base 2025-12-04T08:53:58.7616999Z * [new branch] gh/guilhermeleobas/263/head -> origin/gh/guilhermeleobas/263/head 2025-12-04T08:53:58.7617089Z * [new branch] gh/guilhermeleobas/263/orig -> origin/gh/guilhermeleobas/263/orig 2025-12-04T08:53:58.7617178Z * [new branch] gh/guilhermeleobas/264/base -> origin/gh/guilhermeleobas/264/base 2025-12-04T08:53:58.7617271Z * [new branch] gh/guilhermeleobas/264/head -> origin/gh/guilhermeleobas/264/head 2025-12-04T08:53:58.7617363Z * [new branch] gh/guilhermeleobas/264/orig -> origin/gh/guilhermeleobas/264/orig 2025-12-04T08:53:58.7617497Z * [new branch] gh/guilhermeleobas/265/base -> origin/gh/guilhermeleobas/265/base 2025-12-04T08:53:58.7617588Z * [new branch] gh/guilhermeleobas/265/head -> origin/gh/guilhermeleobas/265/head 2025-12-04T08:53:58.7617680Z * [new branch] gh/guilhermeleobas/265/orig -> origin/gh/guilhermeleobas/265/orig 2025-12-04T08:53:58.7617768Z * [new branch] gh/guilhermeleobas/266/base -> origin/gh/guilhermeleobas/266/base 2025-12-04T08:53:58.7617860Z * [new branch] gh/guilhermeleobas/266/head -> origin/gh/guilhermeleobas/266/head 2025-12-04T08:53:58.7617951Z * [new branch] gh/guilhermeleobas/266/orig -> origin/gh/guilhermeleobas/266/orig 2025-12-04T08:53:58.7618048Z * [new branch] gh/guilhermeleobas/267/base -> origin/gh/guilhermeleobas/267/base 2025-12-04T08:53:58.7618141Z * [new branch] gh/guilhermeleobas/267/head -> origin/gh/guilhermeleobas/267/head 2025-12-04T08:53:58.7618233Z * [new branch] gh/guilhermeleobas/267/orig -> origin/gh/guilhermeleobas/267/orig 2025-12-04T08:53:58.7618322Z * [new branch] gh/hameerabbasi/1/base -> origin/gh/hameerabbasi/1/base 2025-12-04T08:53:58.7618402Z * [new branch] gh/hameerabbasi/1/head -> origin/gh/hameerabbasi/1/head 2025-12-04T08:53:58.7618481Z * [new branch] gh/hameerabbasi/2/base -> origin/gh/hameerabbasi/2/base 2025-12-04T08:53:58.7618560Z * [new branch] gh/hameerabbasi/2/head -> origin/gh/hameerabbasi/2/head 2025-12-04T08:53:58.7618635Z * [new branch] gh/hameerabbasi/2/orig -> origin/gh/hameerabbasi/2/orig 2025-12-04T08:53:58.7618712Z * [new branch] gh/hameerabbasi/3/base -> origin/gh/hameerabbasi/3/base 2025-12-04T08:53:58.7618789Z * [new branch] gh/hameerabbasi/3/head -> origin/gh/hameerabbasi/3/head 2025-12-04T08:53:58.7618862Z * [new branch] gh/hameerabbasi/3/orig -> origin/gh/hameerabbasi/3/orig 2025-12-04T08:53:58.7618939Z * [new branch] gh/hameerabbasi/4/base -> origin/gh/hameerabbasi/4/base 2025-12-04T08:53:58.7619016Z * [new branch] gh/hameerabbasi/4/head -> origin/gh/hameerabbasi/4/head 2025-12-04T08:53:58.7619091Z * [new branch] gh/hameerabbasi/4/orig -> origin/gh/hameerabbasi/4/orig 2025-12-04T08:53:58.7619163Z * [new branch] gh/huydhn/1/next -> origin/gh/huydhn/1/next 2025-12-04T08:53:58.7619242Z * [new branch] gh/huydhn/2/next -> origin/gh/huydhn/2/next 2025-12-04T08:53:58.7619313Z * [new branch] gh/huydhn/3/next -> origin/gh/huydhn/3/next 2025-12-04T08:53:58.7619429Z * [new branch] gh/huydhn/4/next -> origin/gh/huydhn/4/next 2025-12-04T08:53:58.7619503Z * [new branch] gh/huydhn/5/next -> origin/gh/huydhn/5/next 2025-12-04T08:53:58.7619571Z * [new branch] gh/huydhn/6/next -> origin/gh/huydhn/6/next 2025-12-04T08:53:58.7619642Z * [new branch] gh/int3/97/base -> origin/gh/int3/97/base 2025-12-04T08:53:58.7619714Z * [new branch] gh/int3/97/head -> origin/gh/int3/97/head 2025-12-04T08:53:58.7619787Z * [new branch] gh/isuruf/101/base -> origin/gh/isuruf/101/base 2025-12-04T08:53:58.7619860Z * [new branch] gh/isuruf/101/head -> origin/gh/isuruf/101/head 2025-12-04T08:53:58.7619932Z * [new branch] gh/isuruf/146/base -> origin/gh/isuruf/146/base 2025-12-04T08:53:58.7620004Z * [new branch] gh/isuruf/146/head -> origin/gh/isuruf/146/head 2025-12-04T08:53:58.7620084Z * [new branch] gh/isuruf/146/orig -> origin/gh/isuruf/146/orig 2025-12-04T08:53:58.7620156Z * [new branch] gh/isuruf/158/base -> origin/gh/isuruf/158/base 2025-12-04T08:53:58.7620227Z * [new branch] gh/isuruf/158/head -> origin/gh/isuruf/158/head 2025-12-04T08:53:58.7620328Z * [new branch] gh/isuruf/159/base -> origin/gh/isuruf/159/base 2025-12-04T08:53:58.7620397Z * [new branch] gh/isuruf/159/head -> origin/gh/isuruf/159/head 2025-12-04T08:53:58.7620466Z * [new branch] gh/isuruf/160/base -> origin/gh/isuruf/160/base 2025-12-04T08:53:58.7620539Z * [new branch] gh/isuruf/160/head -> origin/gh/isuruf/160/head 2025-12-04T08:53:58.7620611Z * [new branch] gh/isuruf/160/orig -> origin/gh/isuruf/160/orig 2025-12-04T08:53:58.7620680Z * [new branch] gh/isuruf/81/base -> origin/gh/isuruf/81/base 2025-12-04T08:53:58.7620753Z * [new branch] gh/isuruf/81/head -> origin/gh/isuruf/81/head 2025-12-04T08:53:58.7620821Z * [new branch] gh/isuruf/81/orig -> origin/gh/isuruf/81/orig 2025-12-04T08:53:58.7620897Z * [new branch] gh/jamesjwu/176/base -> origin/gh/jamesjwu/176/base 2025-12-04T08:53:58.7620975Z * [new branch] gh/jamesjwu/176/head -> origin/gh/jamesjwu/176/head 2025-12-04T08:53:58.7621048Z * [new branch] gh/jamesjwu/176/orig -> origin/gh/jamesjwu/176/orig 2025-12-04T08:53:58.7621121Z * [new branch] gh/jamesjwu/187/base -> origin/gh/jamesjwu/187/base 2025-12-04T08:53:58.7621194Z * [new branch] gh/jamesjwu/187/head -> origin/gh/jamesjwu/187/head 2025-12-04T08:53:58.7621266Z * [new branch] gh/jamesjwu/187/orig -> origin/gh/jamesjwu/187/orig 2025-12-04T08:53:58.7621338Z * [new branch] gh/jamesjwu/196/base -> origin/gh/jamesjwu/196/base 2025-12-04T08:53:58.7621413Z * [new branch] gh/jamesjwu/196/head -> origin/gh/jamesjwu/196/head 2025-12-04T08:53:58.7621484Z * [new branch] gh/jamesjwu/196/orig -> origin/gh/jamesjwu/196/orig 2025-12-04T08:53:58.7621556Z * [new branch] gh/jamesjwu/198/base -> origin/gh/jamesjwu/198/base 2025-12-04T08:53:58.7621629Z * [new branch] gh/jamesjwu/198/head -> origin/gh/jamesjwu/198/head 2025-12-04T08:53:58.7621701Z * [new branch] gh/jamesjwu/198/orig -> origin/gh/jamesjwu/198/orig 2025-12-04T08:53:58.7621775Z * [new branch] gh/jamesjwu/207/base -> origin/gh/jamesjwu/207/base 2025-12-04T08:53:58.7621847Z * [new branch] gh/jamesjwu/207/head -> origin/gh/jamesjwu/207/head 2025-12-04T08:53:58.7621918Z * [new branch] gh/jamesjwu/207/orig -> origin/gh/jamesjwu/207/orig 2025-12-04T08:53:58.7621990Z * [new branch] gh/jamesjwu/208/base -> origin/gh/jamesjwu/208/base 2025-12-04T08:53:58.7622089Z * [new branch] gh/jamesjwu/208/head -> origin/gh/jamesjwu/208/head 2025-12-04T08:53:58.7622159Z * [new branch] gh/jamesjwu/208/orig -> origin/gh/jamesjwu/208/orig 2025-12-04T08:53:58.7622233Z * [new branch] gh/jamesjwu/52/base -> origin/gh/jamesjwu/52/base 2025-12-04T08:53:58.7622307Z * [new branch] gh/jamesjwu/52/head -> origin/gh/jamesjwu/52/head 2025-12-04T08:53:58.7622378Z * [new branch] gh/jamesjwu/53/base -> origin/gh/jamesjwu/53/base 2025-12-04T08:53:58.7622454Z * [new branch] gh/jamesjwu/53/head -> origin/gh/jamesjwu/53/head 2025-12-04T08:53:58.7622527Z * [new branch] gh/jamesjwu/54/base -> origin/gh/jamesjwu/54/base 2025-12-04T08:53:58.7622600Z * [new branch] gh/jamesjwu/54/head -> origin/gh/jamesjwu/54/head 2025-12-04T08:53:58.7622831Z * [new branch] gh/jamesjwu/55/base -> origin/gh/jamesjwu/55/base 2025-12-04T08:53:58.7622905Z * [new branch] gh/jamesjwu/55/head -> origin/gh/jamesjwu/55/head 2025-12-04T08:53:58.7622977Z * [new branch] gh/jamesjwu/56/base -> origin/gh/jamesjwu/56/base 2025-12-04T08:53:58.7623055Z * [new branch] gh/jamesjwu/56/head -> origin/gh/jamesjwu/56/head 2025-12-04T08:53:58.7623180Z * [new branch] gh/jamesjwu/57/base -> origin/gh/jamesjwu/57/base 2025-12-04T08:53:58.7623273Z * [new branch] gh/jamesjwu/57/head -> origin/gh/jamesjwu/57/head 2025-12-04T08:53:58.7623345Z * [new branch] gh/jamesjwu/58/base -> origin/gh/jamesjwu/58/base 2025-12-04T08:53:58.7623414Z * [new branch] gh/jamesjwu/58/head -> origin/gh/jamesjwu/58/head 2025-12-04T08:53:58.7623488Z * [new branch] gh/jamesjwu/59/base -> origin/gh/jamesjwu/59/base 2025-12-04T08:53:58.7623561Z * [new branch] gh/jamesjwu/59/head -> origin/gh/jamesjwu/59/head 2025-12-04T08:53:58.7623634Z * [new branch] gh/jamesjwu/60/base -> origin/gh/jamesjwu/60/base 2025-12-04T08:53:58.7623708Z * [new branch] gh/jamesjwu/60/head -> origin/gh/jamesjwu/60/head 2025-12-04T08:53:58.7623798Z * [new branch] gh/jamesjwu/61/base -> origin/gh/jamesjwu/61/base 2025-12-04T08:53:58.7623872Z * [new branch] gh/jamesjwu/61/head -> origin/gh/jamesjwu/61/head 2025-12-04T08:53:58.7623949Z * [new branch] gh/jamesjwu/62/base -> origin/gh/jamesjwu/62/base 2025-12-04T08:53:58.7624020Z * [new branch] gh/jamesjwu/62/head -> origin/gh/jamesjwu/62/head 2025-12-04T08:53:58.7624093Z * [new branch] gh/jamesjwu/63/base -> origin/gh/jamesjwu/63/base 2025-12-04T08:53:58.7624171Z * [new branch] gh/jamesjwu/63/head -> origin/gh/jamesjwu/63/head 2025-12-04T08:53:58.7624243Z * [new branch] gh/jamesjwu/64/base -> origin/gh/jamesjwu/64/base 2025-12-04T08:53:58.7624315Z * [new branch] gh/jamesjwu/64/head -> origin/gh/jamesjwu/64/head 2025-12-04T08:53:58.7624385Z * [new branch] gh/jamesjwu/65/base -> origin/gh/jamesjwu/65/base 2025-12-04T08:53:58.7624457Z * [new branch] gh/jamesjwu/65/head -> origin/gh/jamesjwu/65/head 2025-12-04T08:53:58.7624534Z * [new branch] gh/janeyx99/165/base -> origin/gh/janeyx99/165/base 2025-12-04T08:53:58.7624612Z * [new branch] gh/janeyx99/165/head -> origin/gh/janeyx99/165/head 2025-12-04T08:53:58.7624686Z * [new branch] gh/janeyx99/165/orig -> origin/gh/janeyx99/165/orig 2025-12-04T08:53:58.7624758Z * [new branch] gh/janeyx99/201/base -> origin/gh/janeyx99/201/base 2025-12-04T08:53:58.7624834Z * [new branch] gh/janeyx99/201/head -> origin/gh/janeyx99/201/head 2025-12-04T08:53:58.7624906Z * [new branch] gh/janeyx99/201/orig -> origin/gh/janeyx99/201/orig 2025-12-04T08:53:58.7625015Z * [new branch] gh/janeyx99/225/base -> origin/gh/janeyx99/225/base 2025-12-04T08:53:58.7625089Z * [new branch] gh/janeyx99/225/head -> origin/gh/janeyx99/225/head 2025-12-04T08:53:58.7625160Z * [new branch] gh/janeyx99/225/orig -> origin/gh/janeyx99/225/orig 2025-12-04T08:53:58.7625237Z * [new branch] gh/janeyx99/299/base -> origin/gh/janeyx99/299/base 2025-12-04T08:53:58.7625307Z * [new branch] gh/janeyx99/299/head -> origin/gh/janeyx99/299/head 2025-12-04T08:53:58.7625380Z * [new branch] gh/janeyx99/299/orig -> origin/gh/janeyx99/299/orig 2025-12-04T08:53:58.7625458Z * [new branch] gh/janeyx99/302/base -> origin/gh/janeyx99/302/base 2025-12-04T08:53:58.7625530Z * [new branch] gh/janeyx99/302/head -> origin/gh/janeyx99/302/head 2025-12-04T08:53:58.7625605Z * [new branch] gh/janeyx99/303/base -> origin/gh/janeyx99/303/base 2025-12-04T08:53:58.7625679Z * [new branch] gh/janeyx99/303/head -> origin/gh/janeyx99/303/head 2025-12-04T08:53:58.7625751Z * [new branch] gh/janeyx99/305/base -> origin/gh/janeyx99/305/base 2025-12-04T08:53:58.7625825Z * [new branch] gh/janeyx99/305/head -> origin/gh/janeyx99/305/head 2025-12-04T08:53:58.7625930Z * [new branch] gh/janeyx99/306/base -> origin/gh/janeyx99/306/base 2025-12-04T08:53:58.7626003Z * [new branch] gh/janeyx99/306/head -> origin/gh/janeyx99/306/head 2025-12-04T08:53:58.7626074Z * [new branch] gh/janeyx99/314/base -> origin/gh/janeyx99/314/base 2025-12-04T08:53:58.7626145Z * [new branch] gh/janeyx99/314/head -> origin/gh/janeyx99/314/head 2025-12-04T08:53:58.7626216Z * [new branch] gh/janeyx99/314/orig -> origin/gh/janeyx99/314/orig 2025-12-04T08:53:58.7626288Z * [new branch] gh/janeyx99/315/base -> origin/gh/janeyx99/315/base 2025-12-04T08:53:58.7626360Z * [new branch] gh/janeyx99/315/head -> origin/gh/janeyx99/315/head 2025-12-04T08:53:58.7626428Z * [new branch] gh/janeyx99/315/orig -> origin/gh/janeyx99/315/orig 2025-12-04T08:53:58.7626499Z * [new branch] gh/janeyx99/316/base -> origin/gh/janeyx99/316/base 2025-12-04T08:53:58.7626575Z * [new branch] gh/janeyx99/316/head -> origin/gh/janeyx99/316/head 2025-12-04T08:53:58.7626644Z * [new branch] gh/janeyx99/316/orig -> origin/gh/janeyx99/316/orig 2025-12-04T08:53:58.7626716Z * [new branch] gh/janeyx99/317/base -> origin/gh/janeyx99/317/base 2025-12-04T08:53:58.7626832Z * [new branch] gh/janeyx99/317/head -> origin/gh/janeyx99/317/head 2025-12-04T08:53:58.7626905Z * [new branch] gh/janeyx99/317/orig -> origin/gh/janeyx99/317/orig 2025-12-04T08:53:58.7626977Z * [new branch] gh/janeyx99/325/base -> origin/gh/janeyx99/325/base 2025-12-04T08:53:58.7627046Z * [new branch] gh/janeyx99/325/head -> origin/gh/janeyx99/325/head 2025-12-04T08:53:58.7627118Z * [new branch] gh/janeyx99/325/orig -> origin/gh/janeyx99/325/orig 2025-12-04T08:53:58.7627188Z * [new branch] gh/janeyx99/327/base -> origin/gh/janeyx99/327/base 2025-12-04T08:53:58.7627260Z * [new branch] gh/janeyx99/327/head -> origin/gh/janeyx99/327/head 2025-12-04T08:53:58.7627330Z * [new branch] gh/janeyx99/327/orig -> origin/gh/janeyx99/327/orig 2025-12-04T08:53:58.7627402Z * [new branch] gh/janeyx99/328/base -> origin/gh/janeyx99/328/base 2025-12-04T08:53:58.7627472Z * [new branch] gh/janeyx99/328/head -> origin/gh/janeyx99/328/head 2025-12-04T08:53:58.7627543Z * [new branch] gh/janeyx99/328/orig -> origin/gh/janeyx99/328/orig 2025-12-04T08:53:58.7627650Z * [new branch] gh/janeyx99/329/base -> origin/gh/janeyx99/329/base 2025-12-04T08:53:58.7627720Z * [new branch] gh/janeyx99/329/head -> origin/gh/janeyx99/329/head 2025-12-04T08:53:58.7627790Z * [new branch] gh/janeyx99/329/orig -> origin/gh/janeyx99/329/orig 2025-12-04T08:53:58.7627863Z * [new branch] gh/janeyx99/330/base -> origin/gh/janeyx99/330/base 2025-12-04T08:53:58.7627934Z * [new branch] gh/janeyx99/330/head -> origin/gh/janeyx99/330/head 2025-12-04T08:53:58.7628004Z * [new branch] gh/janeyx99/330/orig -> origin/gh/janeyx99/330/orig 2025-12-04T08:53:58.7628074Z * [new branch] gh/janeyx99/331/base -> origin/gh/janeyx99/331/base 2025-12-04T08:53:58.7628144Z * [new branch] gh/janeyx99/331/head -> origin/gh/janeyx99/331/head 2025-12-04T08:53:58.7628217Z * [new branch] gh/janeyx99/331/orig -> origin/gh/janeyx99/331/orig 2025-12-04T08:53:58.7628289Z * [new branch] gh/janeyx99/332/base -> origin/gh/janeyx99/332/base 2025-12-04T08:53:58.7628360Z * [new branch] gh/janeyx99/332/head -> origin/gh/janeyx99/332/head 2025-12-04T08:53:58.7628432Z * [new branch] gh/janeyx99/332/orig -> origin/gh/janeyx99/332/orig 2025-12-04T08:53:58.7628563Z * [new branch] gh/janeyx99/333/base -> origin/gh/janeyx99/333/base 2025-12-04T08:53:58.7628635Z * [new branch] gh/janeyx99/333/head -> origin/gh/janeyx99/333/head 2025-12-04T08:53:58.7628706Z * [new branch] gh/janeyx99/333/orig -> origin/gh/janeyx99/333/orig 2025-12-04T08:53:58.7628776Z * [new branch] gh/janeyx99/88/base -> origin/gh/janeyx99/88/base 2025-12-04T08:53:58.7628846Z * [new branch] gh/janeyx99/88/head -> origin/gh/janeyx99/88/head 2025-12-04T08:53:58.7628917Z * [new branch] gh/janeyx99/88/orig -> origin/gh/janeyx99/88/orig 2025-12-04T08:53:58.7628991Z * [new branch] gh/jansel/360/base -> origin/gh/jansel/360/base 2025-12-04T08:53:58.7629061Z * [new branch] gh/jansel/360/head -> origin/gh/jansel/360/head 2025-12-04T08:53:58.7629132Z * [new branch] gh/jansel/451/base -> origin/gh/jansel/451/base 2025-12-04T08:53:58.7629203Z * [new branch] gh/jansel/451/head -> origin/gh/jansel/451/head 2025-12-04T08:53:58.7629271Z * [new branch] gh/jansel/451/orig -> origin/gh/jansel/451/orig 2025-12-04T08:53:58.7629343Z * [new branch] gh/jansel/462/base -> origin/gh/jansel/462/base 2025-12-04T08:53:58.7629413Z * [new branch] gh/jansel/462/head -> origin/gh/jansel/462/head 2025-12-04T08:53:58.7629482Z * [new branch] gh/jansel/462/orig -> origin/gh/jansel/462/orig 2025-12-04T08:53:58.7629551Z * [new branch] gh/jansel/533/base -> origin/gh/jansel/533/base 2025-12-04T08:53:58.7629620Z * [new branch] gh/jansel/533/head -> origin/gh/jansel/533/head 2025-12-04T08:53:58.7629688Z * [new branch] gh/jansel/533/orig -> origin/gh/jansel/533/orig 2025-12-04T08:53:58.7629759Z * [new branch] gh/jansel/552/base -> origin/gh/jansel/552/base 2025-12-04T08:53:58.7629828Z * [new branch] gh/jansel/552/head -> origin/gh/jansel/552/head 2025-12-04T08:53:58.7629898Z * [new branch] gh/jansel/552/orig -> origin/gh/jansel/552/orig 2025-12-04T08:53:58.7629967Z * [new branch] gh/jansel/553/base -> origin/gh/jansel/553/base 2025-12-04T08:53:58.7630036Z * [new branch] gh/jansel/553/head -> origin/gh/jansel/553/head 2025-12-04T08:53:58.7630105Z * [new branch] gh/jansel/553/orig -> origin/gh/jansel/553/orig 2025-12-04T08:53:58.7630172Z * [new branch] gh/jansel/554/base -> origin/gh/jansel/554/base 2025-12-04T08:53:58.7630741Z * [new branch] gh/jansel/554/head -> origin/gh/jansel/554/head 2025-12-04T08:53:58.7630811Z * [new branch] gh/jansel/554/orig -> origin/gh/jansel/554/orig 2025-12-04T08:53:58.7630881Z * [new branch] gh/jansel/555/base -> origin/gh/jansel/555/base 2025-12-04T08:53:58.7630950Z * [new branch] gh/jansel/555/head -> origin/gh/jansel/555/head 2025-12-04T08:53:58.7631022Z * [new branch] gh/jansel/555/orig -> origin/gh/jansel/555/orig 2025-12-04T08:53:58.7631090Z * [new branch] gh/jansel/556/base -> origin/gh/jansel/556/base 2025-12-04T08:53:58.7631158Z * [new branch] gh/jansel/556/head -> origin/gh/jansel/556/head 2025-12-04T08:53:58.7631228Z * [new branch] gh/jansel/556/orig -> origin/gh/jansel/556/orig 2025-12-04T08:53:58.7631296Z * [new branch] gh/jansel/557/base -> origin/gh/jansel/557/base 2025-12-04T08:53:58.7631365Z * [new branch] gh/jansel/557/head -> origin/gh/jansel/557/head 2025-12-04T08:53:58.7631435Z * [new branch] gh/jansel/557/orig -> origin/gh/jansel/557/orig 2025-12-04T08:53:58.7631504Z * [new branch] gh/jansel/558/base -> origin/gh/jansel/558/base 2025-12-04T08:53:58.7631572Z * [new branch] gh/jansel/558/head -> origin/gh/jansel/558/head 2025-12-04T08:53:58.7631675Z * [new branch] gh/jansel/558/orig -> origin/gh/jansel/558/orig 2025-12-04T08:53:58.7631745Z * [new branch] gh/jansel/559/base -> origin/gh/jansel/559/base 2025-12-04T08:53:58.7631815Z * [new branch] gh/jansel/559/head -> origin/gh/jansel/559/head 2025-12-04T08:53:58.7631883Z * [new branch] gh/jansel/559/orig -> origin/gh/jansel/559/orig 2025-12-04T08:53:58.7631951Z * [new branch] gh/jansel/560/base -> origin/gh/jansel/560/base 2025-12-04T08:53:58.7632022Z * [new branch] gh/jansel/560/head -> origin/gh/jansel/560/head 2025-12-04T08:53:58.7632089Z * [new branch] gh/jansel/560/orig -> origin/gh/jansel/560/orig 2025-12-04T08:53:58.7632157Z * [new branch] gh/jansel/561/base -> origin/gh/jansel/561/base 2025-12-04T08:53:58.7632226Z * [new branch] gh/jansel/561/head -> origin/gh/jansel/561/head 2025-12-04T08:53:58.7632296Z * [new branch] gh/jansel/561/orig -> origin/gh/jansel/561/orig 2025-12-04T08:53:58.7632364Z * [new branch] gh/jansel/562/base -> origin/gh/jansel/562/base 2025-12-04T08:53:58.7632434Z * [new branch] gh/jansel/562/head -> origin/gh/jansel/562/head 2025-12-04T08:53:58.7632502Z * [new branch] gh/jansel/562/orig -> origin/gh/jansel/562/orig 2025-12-04T08:53:58.7632570Z * [new branch] gh/jansel/563/base -> origin/gh/jansel/563/base 2025-12-04T08:53:58.7632641Z * [new branch] gh/jansel/563/head -> origin/gh/jansel/563/head 2025-12-04T08:53:58.7632709Z * [new branch] gh/jansel/563/orig -> origin/gh/jansel/563/orig 2025-12-04T08:53:58.7632777Z * [new branch] gh/jansel/564/base -> origin/gh/jansel/564/base 2025-12-04T08:53:58.7632847Z * [new branch] gh/jansel/564/head -> origin/gh/jansel/564/head 2025-12-04T08:53:58.7632916Z * [new branch] gh/jansel/564/orig -> origin/gh/jansel/564/orig 2025-12-04T08:53:58.7632984Z * [new branch] gh/jansel/565/base -> origin/gh/jansel/565/base 2025-12-04T08:53:58.7633053Z * [new branch] gh/jansel/565/head -> origin/gh/jansel/565/head 2025-12-04T08:53:58.7633121Z * [new branch] gh/jansel/565/orig -> origin/gh/jansel/565/orig 2025-12-04T08:53:58.7633188Z * [new branch] gh/jansel/566/base -> origin/gh/jansel/566/base 2025-12-04T08:53:58.7633285Z * [new branch] gh/jansel/566/head -> origin/gh/jansel/566/head 2025-12-04T08:53:58.7633353Z * [new branch] gh/jansel/566/orig -> origin/gh/jansel/566/orig 2025-12-04T08:53:58.7633421Z * [new branch] gh/jansel/567/base -> origin/gh/jansel/567/base 2025-12-04T08:53:58.7633489Z * [new branch] gh/jansel/567/head -> origin/gh/jansel/567/head 2025-12-04T08:53:58.7633558Z * [new branch] gh/jansel/567/orig -> origin/gh/jansel/567/orig 2025-12-04T08:53:58.7633628Z * [new branch] gh/jansel/568/base -> origin/gh/jansel/568/base 2025-12-04T08:53:58.7633696Z * [new branch] gh/jansel/568/head -> origin/gh/jansel/568/head 2025-12-04T08:53:58.7633763Z * [new branch] gh/jansel/568/orig -> origin/gh/jansel/568/orig 2025-12-04T08:53:58.7633832Z * [new branch] gh/jansel/569/base -> origin/gh/jansel/569/base 2025-12-04T08:53:58.7633902Z * [new branch] gh/jansel/569/head -> origin/gh/jansel/569/head 2025-12-04T08:53:58.7633971Z * [new branch] gh/jansel/569/orig -> origin/gh/jansel/569/orig 2025-12-04T08:53:58.7634041Z * [new branch] gh/jansel/570/base -> origin/gh/jansel/570/base 2025-12-04T08:53:58.7634109Z * [new branch] gh/jansel/570/head -> origin/gh/jansel/570/head 2025-12-04T08:53:58.7634201Z * [new branch] gh/jansel/570/orig -> origin/gh/jansel/570/orig 2025-12-04T08:53:58.7634272Z * [new branch] gh/jansel/571/base -> origin/gh/jansel/571/base 2025-12-04T08:53:58.7634339Z * [new branch] gh/jansel/571/head -> origin/gh/jansel/571/head 2025-12-04T08:53:58.7634407Z * [new branch] gh/jansel/571/orig -> origin/gh/jansel/571/orig 2025-12-04T08:53:58.7634476Z * [new branch] gh/jansel/572/base -> origin/gh/jansel/572/base 2025-12-04T08:53:58.7634546Z * [new branch] gh/jansel/572/head -> origin/gh/jansel/572/head 2025-12-04T08:53:58.7634614Z * [new branch] gh/jansel/572/orig -> origin/gh/jansel/572/orig 2025-12-04T08:53:58.7634827Z * [new branch] gh/jansel/573/base -> origin/gh/jansel/573/base 2025-12-04T08:53:58.7634895Z * [new branch] gh/jansel/573/head -> origin/gh/jansel/573/head 2025-12-04T08:53:58.7634965Z * [new branch] gh/jansel/573/orig -> origin/gh/jansel/573/orig 2025-12-04T08:53:58.7635035Z * [new branch] gh/jansel/574/base -> origin/gh/jansel/574/base 2025-12-04T08:53:58.7635102Z * [new branch] gh/jansel/574/head -> origin/gh/jansel/574/head 2025-12-04T08:53:58.7635170Z * [new branch] gh/jansel/574/orig -> origin/gh/jansel/574/orig 2025-12-04T08:53:58.7635238Z * [new branch] gh/jansel/575/base -> origin/gh/jansel/575/base 2025-12-04T08:53:58.7635305Z * [new branch] gh/jansel/575/head -> origin/gh/jansel/575/head 2025-12-04T08:53:58.7635375Z * [new branch] gh/jansel/575/orig -> origin/gh/jansel/575/orig 2025-12-04T08:53:58.7635445Z * [new branch] gh/jansel/576/base -> origin/gh/jansel/576/base 2025-12-04T08:53:58.7635513Z * [new branch] gh/jansel/576/head -> origin/gh/jansel/576/head 2025-12-04T08:53:58.7635584Z * [new branch] gh/jansel/576/orig -> origin/gh/jansel/576/orig 2025-12-04T08:53:58.7635667Z * [new branch] gh/jbschlosser/247/base -> origin/gh/jbschlosser/247/base 2025-12-04T08:53:58.7635747Z * [new branch] gh/jbschlosser/247/head -> origin/gh/jbschlosser/247/head 2025-12-04T08:53:58.7635826Z * [new branch] gh/jbschlosser/247/orig -> origin/gh/jbschlosser/247/orig 2025-12-04T08:53:58.7635903Z * [new branch] gh/jbschlosser/250/base -> origin/gh/jbschlosser/250/base 2025-12-04T08:53:58.7636011Z * [new branch] gh/jbschlosser/250/head -> origin/gh/jbschlosser/250/head 2025-12-04T08:53:58.7636088Z * [new branch] gh/jbschlosser/250/orig -> origin/gh/jbschlosser/250/orig 2025-12-04T08:53:58.7636162Z * [new branch] gh/jerryzh168/1/base -> origin/gh/jerryzh168/1/base 2025-12-04T08:53:58.7636235Z * [new branch] gh/jerryzh168/1/head -> origin/gh/jerryzh168/1/head 2025-12-04T08:53:58.7636310Z * [new branch] gh/jerryzh168/1/orig -> origin/gh/jerryzh168/1/orig 2025-12-04T08:53:58.7636382Z * [new branch] gh/jiayisunx/59/base -> origin/gh/jiayisunx/59/base 2025-12-04T08:53:58.7636454Z * [new branch] gh/jiayisunx/59/head -> origin/gh/jiayisunx/59/head 2025-12-04T08:53:58.7636525Z * [new branch] gh/jiayisunx/59/orig -> origin/gh/jiayisunx/59/orig 2025-12-04T08:53:58.7636596Z * [new branch] gh/jiayisunx/61/base -> origin/gh/jiayisunx/61/base 2025-12-04T08:53:58.7636788Z * [new branch] gh/jiayisunx/61/head -> origin/gh/jiayisunx/61/head 2025-12-04T08:53:58.7636860Z * [new branch] gh/jiayisunx/61/orig -> origin/gh/jiayisunx/61/orig 2025-12-04T08:53:58.7636932Z * [new branch] gh/jiayisunx/68/base -> origin/gh/jiayisunx/68/base 2025-12-04T08:53:58.7637003Z * [new branch] gh/jiayisunx/68/head -> origin/gh/jiayisunx/68/head 2025-12-04T08:53:58.7637115Z * [new branch] gh/jiayisunx/68/orig -> origin/gh/jiayisunx/68/orig 2025-12-04T08:53:58.7637187Z * [new branch] gh/jiayisunx/77/base -> origin/gh/jiayisunx/77/base 2025-12-04T08:53:58.7637260Z * [new branch] gh/jiayisunx/77/head -> origin/gh/jiayisunx/77/head 2025-12-04T08:53:58.7637331Z * [new branch] gh/jiayisunx/77/orig -> origin/gh/jiayisunx/77/orig 2025-12-04T08:53:58.7637403Z * [new branch] gh/jiayisunx/78/base -> origin/gh/jiayisunx/78/base 2025-12-04T08:53:58.7637480Z * [new branch] gh/jiayisunx/78/head -> origin/gh/jiayisunx/78/head 2025-12-04T08:53:58.7637551Z * [new branch] gh/jiayisunx/78/orig -> origin/gh/jiayisunx/78/orig 2025-12-04T08:53:58.7637622Z * [new branch] gh/jiayisunx/79/base -> origin/gh/jiayisunx/79/base 2025-12-04T08:53:58.7637695Z * [new branch] gh/jiayisunx/79/head -> origin/gh/jiayisunx/79/head 2025-12-04T08:53:58.7637767Z * [new branch] gh/jiayisunx/79/orig -> origin/gh/jiayisunx/79/orig 2025-12-04T08:53:58.7637839Z * [new branch] gh/jiayisunx/82/base -> origin/gh/jiayisunx/82/base 2025-12-04T08:53:58.7637911Z * [new branch] gh/jiayisunx/82/head -> origin/gh/jiayisunx/82/head 2025-12-04T08:53:58.7637982Z * [new branch] gh/jiayisunx/82/orig -> origin/gh/jiayisunx/82/orig 2025-12-04T08:53:58.7638054Z * [new branch] gh/jiayisunx/83/base -> origin/gh/jiayisunx/83/base 2025-12-04T08:53:58.7638128Z * [new branch] gh/jiayisunx/83/head -> origin/gh/jiayisunx/83/head 2025-12-04T08:53:58.7638199Z * [new branch] gh/jiayisunx/83/orig -> origin/gh/jiayisunx/83/orig 2025-12-04T08:53:58.7638270Z * [new branch] gh/jiayisunx/84/base -> origin/gh/jiayisunx/84/base 2025-12-04T08:53:58.7638343Z * [new branch] gh/jiayisunx/84/head -> origin/gh/jiayisunx/84/head 2025-12-04T08:53:58.7638414Z * [new branch] gh/jiayisunx/84/orig -> origin/gh/jiayisunx/84/orig 2025-12-04T08:53:58.7638485Z * [new branch] gh/jiayisunx/85/base -> origin/gh/jiayisunx/85/base 2025-12-04T08:53:58.7638557Z * [new branch] gh/jiayisunx/85/head -> origin/gh/jiayisunx/85/head 2025-12-04T08:53:58.7638628Z * [new branch] gh/jiayisunx/85/orig -> origin/gh/jiayisunx/85/orig 2025-12-04T08:53:58.7638700Z * [new branch] gh/jiayisunx/86/base -> origin/gh/jiayisunx/86/base 2025-12-04T08:53:58.7638809Z * [new branch] gh/jiayisunx/86/head -> origin/gh/jiayisunx/86/head 2025-12-04T08:53:58.7638881Z * [new branch] gh/jiayisunx/86/orig -> origin/gh/jiayisunx/86/orig 2025-12-04T08:53:58.7638955Z * [new branch] gh/jiayisunx/87/base -> origin/gh/jiayisunx/87/base 2025-12-04T08:53:58.7639027Z * [new branch] gh/jiayisunx/87/head -> origin/gh/jiayisunx/87/head 2025-12-04T08:53:58.7639098Z * [new branch] gh/jiayisunx/87/orig -> origin/gh/jiayisunx/87/orig 2025-12-04T08:53:58.7639170Z * [new branch] gh/jiayisunx/88/base -> origin/gh/jiayisunx/88/base 2025-12-04T08:53:58.7639242Z * [new branch] gh/jiayisunx/88/head -> origin/gh/jiayisunx/88/head 2025-12-04T08:53:58.7639314Z * [new branch] gh/jiayisunx/88/orig -> origin/gh/jiayisunx/88/orig 2025-12-04T08:53:58.7639387Z * [new branch] gh/jiayisunx/89/base -> origin/gh/jiayisunx/89/base 2025-12-04T08:53:58.7639459Z * [new branch] gh/jiayisunx/89/head -> origin/gh/jiayisunx/89/head 2025-12-04T08:53:58.7639530Z * [new branch] gh/jiayisunx/89/orig -> origin/gh/jiayisunx/89/orig 2025-12-04T08:53:58.7639605Z * [new branch] gh/jiayisunx/90/base -> origin/gh/jiayisunx/90/base 2025-12-04T08:53:58.7639708Z * [new branch] gh/jiayisunx/90/head -> origin/gh/jiayisunx/90/head 2025-12-04T08:53:58.7639780Z * [new branch] gh/jiayisunx/90/orig -> origin/gh/jiayisunx/90/orig 2025-12-04T08:53:58.7639861Z * [new branch] gh/jjwu@meta.com/1/base -> origin/gh/jjwu@meta.com/1/base 2025-12-04T08:53:58.7639937Z * [new branch] gh/jjwu@meta.com/1/head -> origin/gh/jjwu@meta.com/1/head 2025-12-04T08:53:58.7640007Z * [new branch] gh/jturney/1/base -> origin/gh/jturney/1/base 2025-12-04T08:53:58.7640078Z * [new branch] gh/jturney/1/head -> origin/gh/jturney/1/head 2025-12-04T08:53:58.7640150Z * [new branch] gh/jturney/1/orig -> origin/gh/jturney/1/orig 2025-12-04T08:53:58.7640219Z * [new branch] gh/jturney/2/base -> origin/gh/jturney/2/base 2025-12-04T08:53:58.7640288Z * [new branch] gh/jturney/2/head -> origin/gh/jturney/2/head 2025-12-04T08:53:58.7640357Z * [new branch] gh/jturney/2/orig -> origin/gh/jturney/2/orig 2025-12-04T08:53:58.7640437Z * [new branch] gh/karthickai/10/base -> origin/gh/karthickai/10/base 2025-12-04T08:53:58.7640514Z * [new branch] gh/karthickai/10/head -> origin/gh/karthickai/10/head 2025-12-04T08:53:58.7640590Z * [new branch] gh/karthickai/10/orig -> origin/gh/karthickai/10/orig 2025-12-04T08:53:58.7640668Z * [new branch] gh/karthickai/11/base -> origin/gh/karthickai/11/base 2025-12-04T08:53:58.7640743Z * [new branch] gh/karthickai/11/head -> origin/gh/karthickai/11/head 2025-12-04T08:53:58.7640817Z * [new branch] gh/karthickai/11/orig -> origin/gh/karthickai/11/orig 2025-12-04T08:53:58.7640893Z * [new branch] gh/karthickai/12/base -> origin/gh/karthickai/12/base 2025-12-04T08:53:58.7640967Z * [new branch] gh/karthickai/12/head -> origin/gh/karthickai/12/head 2025-12-04T08:53:58.7641043Z * [new branch] gh/karthickai/12/orig -> origin/gh/karthickai/12/orig 2025-12-04T08:53:58.7641121Z * [new branch] gh/karthickai/13/base -> origin/gh/karthickai/13/base 2025-12-04T08:53:58.7641197Z * [new branch] gh/karthickai/13/head -> origin/gh/karthickai/13/head 2025-12-04T08:53:58.7641272Z * [new branch] gh/karthickai/13/orig -> origin/gh/karthickai/13/orig 2025-12-04T08:53:58.7641348Z * [new branch] gh/karthickai/14/base -> origin/gh/karthickai/14/base 2025-12-04T08:53:58.7641422Z * [new branch] gh/karthickai/14/head -> origin/gh/karthickai/14/head 2025-12-04T08:53:58.7641533Z * [new branch] gh/karthickai/14/orig -> origin/gh/karthickai/14/orig 2025-12-04T08:53:58.7641609Z * [new branch] gh/karthickai/15/base -> origin/gh/karthickai/15/base 2025-12-04T08:53:58.7641683Z * [new branch] gh/karthickai/15/head -> origin/gh/karthickai/15/head 2025-12-04T08:53:58.7641759Z * [new branch] gh/karthickai/15/orig -> origin/gh/karthickai/15/orig 2025-12-04T08:53:58.7641836Z * [new branch] gh/karthickai/16/base -> origin/gh/karthickai/16/base 2025-12-04T08:53:58.7641910Z * [new branch] gh/karthickai/16/head -> origin/gh/karthickai/16/head 2025-12-04T08:53:58.7641983Z * [new branch] gh/karthickai/16/orig -> origin/gh/karthickai/16/orig 2025-12-04T08:53:58.7642057Z * [new branch] gh/karthickai/17/base -> origin/gh/karthickai/17/base 2025-12-04T08:53:58.7642132Z * [new branch] gh/karthickai/17/head -> origin/gh/karthickai/17/head 2025-12-04T08:53:58.7642207Z * [new branch] gh/karthickai/17/orig -> origin/gh/karthickai/17/orig 2025-12-04T08:53:58.7642281Z * [new branch] gh/karthickai/18/base -> origin/gh/karthickai/18/base 2025-12-04T08:53:58.7642356Z * [new branch] gh/karthickai/18/head -> origin/gh/karthickai/18/head 2025-12-04T08:53:58.7642459Z * [new branch] gh/karthickai/18/orig -> origin/gh/karthickai/18/orig 2025-12-04T08:53:58.7642534Z * [new branch] gh/karthickai/19/base -> origin/gh/karthickai/19/base 2025-12-04T08:53:58.7642607Z * [new branch] gh/karthickai/19/head -> origin/gh/karthickai/19/head 2025-12-04T08:53:58.7642682Z * [new branch] gh/karthickai/19/orig -> origin/gh/karthickai/19/orig 2025-12-04T08:53:58.7642757Z * [new branch] gh/karthickai/20/base -> origin/gh/karthickai/20/base 2025-12-04T08:53:58.7642832Z * [new branch] gh/karthickai/20/head -> origin/gh/karthickai/20/head 2025-12-04T08:53:58.7642911Z * [new branch] gh/karthickai/20/orig -> origin/gh/karthickai/20/orig 2025-12-04T08:53:58.7642986Z * [new branch] gh/karthickai/21/base -> origin/gh/karthickai/21/base 2025-12-04T08:53:58.7643060Z * [new branch] gh/karthickai/21/head -> origin/gh/karthickai/21/head 2025-12-04T08:53:58.7643141Z * [new branch] gh/karthickai/21/orig -> origin/gh/karthickai/21/orig 2025-12-04T08:53:58.7643215Z * [new branch] gh/karthickai/22/base -> origin/gh/karthickai/22/base 2025-12-04T08:53:58.7643290Z * [new branch] gh/karthickai/22/head -> origin/gh/karthickai/22/head 2025-12-04T08:53:58.7643368Z * [new branch] gh/karthickai/22/orig -> origin/gh/karthickai/22/orig 2025-12-04T08:53:58.7643442Z * [new branch] gh/karthickai/23/base -> origin/gh/karthickai/23/base 2025-12-04T08:53:58.7643519Z * [new branch] gh/karthickai/23/head -> origin/gh/karthickai/23/head 2025-12-04T08:53:58.7643598Z * [new branch] gh/karthickai/23/orig -> origin/gh/karthickai/23/orig 2025-12-04T08:53:58.7643673Z * [new branch] gh/karthickai/24/base -> origin/gh/karthickai/24/base 2025-12-04T08:53:58.7643751Z * [new branch] gh/karthickai/24/head -> origin/gh/karthickai/24/head 2025-12-04T08:53:58.7643825Z * [new branch] gh/karthickai/24/orig -> origin/gh/karthickai/24/orig 2025-12-04T08:53:58.7643900Z * [new branch] gh/karthickai/25/base -> origin/gh/karthickai/25/base 2025-12-04T08:53:58.7643978Z * [new branch] gh/karthickai/25/head -> origin/gh/karthickai/25/head 2025-12-04T08:53:58.7644051Z * [new branch] gh/karthickai/25/orig -> origin/gh/karthickai/25/orig 2025-12-04T08:53:58.7644126Z * [new branch] gh/karthickai/26/base -> origin/gh/karthickai/26/base 2025-12-04T08:53:58.7644229Z * [new branch] gh/karthickai/26/head -> origin/gh/karthickai/26/head 2025-12-04T08:53:58.7644305Z * [new branch] gh/karthickai/26/orig -> origin/gh/karthickai/26/orig 2025-12-04T08:53:58.7644381Z * [new branch] gh/karthickai/6/base -> origin/gh/karthickai/6/base 2025-12-04T08:53:58.7644460Z * [new branch] gh/karthickai/6/head -> origin/gh/karthickai/6/head 2025-12-04T08:53:58.7644534Z * [new branch] gh/karthickai/6/orig -> origin/gh/karthickai/6/orig 2025-12-04T08:53:58.7644607Z * [new branch] gh/krocki/1/base -> origin/gh/krocki/1/base 2025-12-04T08:53:58.7644678Z * [new branch] gh/krocki/1/head -> origin/gh/krocki/1/head 2025-12-04T08:53:58.7644746Z * [new branch] gh/krocki/1/orig -> origin/gh/krocki/1/orig 2025-12-04T08:53:58.7644814Z * [new branch] gh/krocki/2/base -> origin/gh/krocki/2/base 2025-12-04T08:53:58.7644889Z * [new branch] gh/krocki/2/head -> origin/gh/krocki/2/head 2025-12-04T08:53:58.7644956Z * [new branch] gh/krocki/2/orig -> origin/gh/krocki/2/orig 2025-12-04T08:53:58.7645037Z * [new branch] gh/kurtamohler/60/base -> origin/gh/kurtamohler/60/base 2025-12-04T08:53:58.7645147Z * [new branch] gh/kurtamohler/60/head -> origin/gh/kurtamohler/60/head 2025-12-04T08:53:58.7645225Z * [new branch] gh/kurtamohler/60/orig -> origin/gh/kurtamohler/60/orig 2025-12-04T08:53:58.7645306Z * [new branch] gh/kurtamohler/61/base -> origin/gh/kurtamohler/61/base 2025-12-04T08:53:58.7645382Z * [new branch] gh/kurtamohler/61/head -> origin/gh/kurtamohler/61/head 2025-12-04T08:53:58.7645457Z * [new branch] gh/kurtamohler/61/orig -> origin/gh/kurtamohler/61/orig 2025-12-04T08:53:58.7645535Z * [new branch] gh/kurtamohler/62/base -> origin/gh/kurtamohler/62/base 2025-12-04T08:53:58.7645613Z * [new branch] gh/kurtamohler/62/head -> origin/gh/kurtamohler/62/head 2025-12-04T08:53:58.7645687Z * [new branch] gh/kurtamohler/62/orig -> origin/gh/kurtamohler/62/orig 2025-12-04T08:53:58.7645766Z * [new branch] gh/kurtamohler/63/base -> origin/gh/kurtamohler/63/base 2025-12-04T08:53:58.7645842Z * [new branch] gh/kurtamohler/63/head -> origin/gh/kurtamohler/63/head 2025-12-04T08:53:58.7645918Z * [new branch] gh/kurtamohler/63/orig -> origin/gh/kurtamohler/63/orig 2025-12-04T08:53:58.7645996Z * [new branch] gh/kurtamohler/64/base -> origin/gh/kurtamohler/64/base 2025-12-04T08:53:58.7646073Z * [new branch] gh/kurtamohler/64/head -> origin/gh/kurtamohler/64/head 2025-12-04T08:53:58.7646149Z * [new branch] gh/kurtamohler/64/orig -> origin/gh/kurtamohler/64/orig 2025-12-04T08:53:58.7646228Z * [new branch] gh/kurtamohler/65/base -> origin/gh/kurtamohler/65/base 2025-12-04T08:53:58.7646306Z * [new branch] gh/kurtamohler/65/head -> origin/gh/kurtamohler/65/head 2025-12-04T08:53:58.7646382Z * [new branch] gh/kurtamohler/65/orig -> origin/gh/kurtamohler/65/orig 2025-12-04T08:53:58.7646461Z * [new branch] gh/kurtamohler/66/base -> origin/gh/kurtamohler/66/base 2025-12-04T08:53:58.7646540Z * [new branch] gh/kurtamohler/66/head -> origin/gh/kurtamohler/66/head 2025-12-04T08:53:58.7646617Z * [new branch] gh/kurtamohler/66/orig -> origin/gh/kurtamohler/66/orig 2025-12-04T08:53:58.7646695Z * [new branch] gh/kurtamohler/67/base -> origin/gh/kurtamohler/67/base 2025-12-04T08:53:58.7646814Z * [new branch] gh/kurtamohler/67/head -> origin/gh/kurtamohler/67/head 2025-12-04T08:53:58.7646896Z * [new branch] gh/kurtamohler/67/orig -> origin/gh/kurtamohler/67/orig 2025-12-04T08:53:58.7647019Z * [new branch] gh/kwen2501/130/base -> origin/gh/kwen2501/130/base 2025-12-04T08:53:58.7647091Z * [new branch] gh/kwen2501/130/head -> origin/gh/kwen2501/130/head 2025-12-04T08:53:58.7647162Z * [new branch] gh/kwen2501/130/orig -> origin/gh/kwen2501/130/orig 2025-12-04T08:53:58.7647233Z * [new branch] gh/kwen2501/170/base -> origin/gh/kwen2501/170/base 2025-12-04T08:53:58.7647306Z * [new branch] gh/kwen2501/170/head -> origin/gh/kwen2501/170/head 2025-12-04T08:53:58.7647380Z * [new branch] gh/kwen2501/187/base -> origin/gh/kwen2501/187/base 2025-12-04T08:53:58.7647450Z * [new branch] gh/kwen2501/187/head -> origin/gh/kwen2501/187/head 2025-12-04T08:53:58.7647521Z * [new branch] gh/kwen2501/187/orig -> origin/gh/kwen2501/187/orig 2025-12-04T08:53:58.7647597Z * [new branch] gh/kwen2501/188/base -> origin/gh/kwen2501/188/base 2025-12-04T08:53:58.7647668Z * [new branch] gh/kwen2501/188/head -> origin/gh/kwen2501/188/head 2025-12-04T08:53:58.7647740Z * [new branch] gh/kwen2501/188/orig -> origin/gh/kwen2501/188/orig 2025-12-04T08:53:58.7647815Z * [new branch] gh/kwen2501/211/base -> origin/gh/kwen2501/211/base 2025-12-04T08:53:58.7647886Z * [new branch] gh/kwen2501/211/head -> origin/gh/kwen2501/211/head 2025-12-04T08:53:58.7647996Z * [new branch] gh/kwen2501/224/base -> origin/gh/kwen2501/224/base 2025-12-04T08:53:58.7648069Z * [new branch] gh/kwen2501/224/head -> origin/gh/kwen2501/224/head 2025-12-04T08:53:58.7648140Z * [new branch] gh/kwen2501/224/orig -> origin/gh/kwen2501/224/orig 2025-12-04T08:53:58.7648210Z * [new branch] gh/kwen2501/228/base -> origin/gh/kwen2501/228/base 2025-12-04T08:53:58.7648283Z * [new branch] gh/kwen2501/228/head -> origin/gh/kwen2501/228/head 2025-12-04T08:53:58.7648355Z * [new branch] gh/kwen2501/228/orig -> origin/gh/kwen2501/228/orig 2025-12-04T08:53:58.7648426Z * [new branch] gh/kwen2501/234/base -> origin/gh/kwen2501/234/base 2025-12-04T08:53:58.7648499Z * [new branch] gh/kwen2501/234/head -> origin/gh/kwen2501/234/head 2025-12-04T08:53:58.7648571Z * [new branch] gh/kwen2501/234/orig -> origin/gh/kwen2501/234/orig 2025-12-04T08:53:58.7648645Z * [new branch] gh/kwen2501/235/base -> origin/gh/kwen2501/235/base 2025-12-04T08:53:58.7648716Z * [new branch] gh/kwen2501/235/head -> origin/gh/kwen2501/235/head 2025-12-04T08:53:58.7648786Z * [new branch] gh/kwen2501/235/orig -> origin/gh/kwen2501/235/orig 2025-12-04T08:53:58.7648860Z * [new branch] gh/kwen2501/236/base -> origin/gh/kwen2501/236/base 2025-12-04T08:53:58.7648931Z * [new branch] gh/kwen2501/236/head -> origin/gh/kwen2501/236/head 2025-12-04T08:53:58.7649004Z * [new branch] gh/kwen2501/236/orig -> origin/gh/kwen2501/236/orig 2025-12-04T08:53:58.7649077Z * [new branch] gh/kwen2501/237/base -> origin/gh/kwen2501/237/base 2025-12-04T08:53:58.7649148Z * [new branch] gh/kwen2501/237/head -> origin/gh/kwen2501/237/head 2025-12-04T08:53:58.7649221Z * [new branch] gh/kwen2501/237/orig -> origin/gh/kwen2501/237/orig 2025-12-04T08:53:58.7649295Z * [new branch] gh/kwen2501/238/base -> origin/gh/kwen2501/238/base 2025-12-04T08:53:58.7649364Z * [new branch] gh/kwen2501/238/head -> origin/gh/kwen2501/238/head 2025-12-04T08:53:58.7649433Z * [new branch] gh/kwen2501/238/orig -> origin/gh/kwen2501/238/orig 2025-12-04T08:53:58.7649508Z * [new branch] gh/kwen2501/240/base -> origin/gh/kwen2501/240/base 2025-12-04T08:53:58.7649578Z * [new branch] gh/kwen2501/240/head -> origin/gh/kwen2501/240/head 2025-12-04T08:53:58.7649676Z * [new branch] gh/kwen2501/240/orig -> origin/gh/kwen2501/240/orig 2025-12-04T08:53:58.7649750Z * [new branch] gh/kwen2501/241/base -> origin/gh/kwen2501/241/base 2025-12-04T08:53:58.7649821Z * [new branch] gh/kwen2501/241/head -> origin/gh/kwen2501/241/head 2025-12-04T08:53:58.7649893Z * [new branch] gh/kwen2501/241/orig -> origin/gh/kwen2501/241/orig 2025-12-04T08:53:58.7649968Z * [new branch] gh/kwen2501/247/base -> origin/gh/kwen2501/247/base 2025-12-04T08:53:58.7650038Z * [new branch] gh/kwen2501/247/head -> origin/gh/kwen2501/247/head 2025-12-04T08:53:58.7650112Z * [new branch] gh/kwen2501/247/orig -> origin/gh/kwen2501/247/orig 2025-12-04T08:53:58.7650183Z * [new branch] gh/kwen2501/252/base -> origin/gh/kwen2501/252/base 2025-12-04T08:53:58.7650253Z * [new branch] gh/kwen2501/252/head -> origin/gh/kwen2501/252/head 2025-12-04T08:53:58.7650326Z * [new branch] gh/kwen2501/252/orig -> origin/gh/kwen2501/252/orig 2025-12-04T08:53:58.7650395Z * [new branch] gh/kwen2501/259/base -> origin/gh/kwen2501/259/base 2025-12-04T08:53:58.7650465Z * [new branch] gh/kwen2501/259/head -> origin/gh/kwen2501/259/head 2025-12-04T08:53:58.7650567Z * [new branch] gh/kwen2501/259/orig -> origin/gh/kwen2501/259/orig 2025-12-04T08:53:58.7650638Z * [new branch] gh/kwen2501/260/base -> origin/gh/kwen2501/260/base 2025-12-04T08:53:58.7650710Z * [new branch] gh/kwen2501/260/head -> origin/gh/kwen2501/260/head 2025-12-04T08:53:58.7650782Z * [new branch] gh/kwen2501/260/orig -> origin/gh/kwen2501/260/orig 2025-12-04T08:53:58.7650854Z * [new branch] gh/kwen2501/268/base -> origin/gh/kwen2501/268/base 2025-12-04T08:53:58.7650925Z * [new branch] gh/kwen2501/268/head -> origin/gh/kwen2501/268/head 2025-12-04T08:53:58.7651000Z * [new branch] gh/kwen2501/268/orig -> origin/gh/kwen2501/268/orig 2025-12-04T08:53:58.7651071Z * [new branch] gh/kwen2501/269/base -> origin/gh/kwen2501/269/base 2025-12-04T08:53:58.7651142Z * [new branch] gh/kwen2501/269/head -> origin/gh/kwen2501/269/head 2025-12-04T08:53:58.7651216Z * [new branch] gh/kwen2501/269/orig -> origin/gh/kwen2501/269/orig 2025-12-04T08:53:58.7651286Z * [new branch] gh/kwen2501/270/base -> origin/gh/kwen2501/270/base 2025-12-04T08:53:58.7651357Z * [new branch] gh/kwen2501/270/head -> origin/gh/kwen2501/270/head 2025-12-04T08:53:58.7651427Z * [new branch] gh/kwen2501/270/orig -> origin/gh/kwen2501/270/orig 2025-12-04T08:53:58.7651499Z * [new branch] gh/kwen2501/271/base -> origin/gh/kwen2501/271/base 2025-12-04T08:53:58.7651572Z * [new branch] gh/kwen2501/271/head -> origin/gh/kwen2501/271/head 2025-12-04T08:53:58.7651642Z * [new branch] gh/kwen2501/271/orig -> origin/gh/kwen2501/271/orig 2025-12-04T08:53:58.7651712Z * [new branch] gh/kwen2501/274/base -> origin/gh/kwen2501/274/base 2025-12-04T08:53:58.7651784Z * [new branch] gh/kwen2501/274/head -> origin/gh/kwen2501/274/head 2025-12-04T08:53:58.7651855Z * [new branch] gh/kwen2501/274/orig -> origin/gh/kwen2501/274/orig 2025-12-04T08:53:58.7651926Z * [new branch] gh/kwen2501/275/base -> origin/gh/kwen2501/275/base 2025-12-04T08:53:58.7651998Z * [new branch] gh/kwen2501/275/head -> origin/gh/kwen2501/275/head 2025-12-04T08:53:58.7652069Z * [new branch] gh/kwen2501/275/orig -> origin/gh/kwen2501/275/orig 2025-12-04T08:53:58.7652139Z * [new branch] gh/kwen2501/276/base -> origin/gh/kwen2501/276/base 2025-12-04T08:53:58.7652214Z * [new branch] gh/kwen2501/276/head -> origin/gh/kwen2501/276/head 2025-12-04T08:53:58.7652315Z * [new branch] gh/kwen2501/276/orig -> origin/gh/kwen2501/276/orig 2025-12-04T08:53:58.7652385Z * [new branch] gh/kwen2501/277/base -> origin/gh/kwen2501/277/base 2025-12-04T08:53:58.7652457Z * [new branch] gh/kwen2501/277/head -> origin/gh/kwen2501/277/head 2025-12-04T08:53:58.7652530Z * [new branch] gh/kwen2501/277/orig -> origin/gh/kwen2501/277/orig 2025-12-04T08:53:58.7652599Z * [new branch] gh/kwen2501/278/base -> origin/gh/kwen2501/278/base 2025-12-04T08:53:58.7652672Z * [new branch] gh/kwen2501/278/head -> origin/gh/kwen2501/278/head 2025-12-04T08:53:58.7652742Z * [new branch] gh/kwen2501/278/orig -> origin/gh/kwen2501/278/orig 2025-12-04T08:53:58.7652812Z * [new branch] gh/kwen2501/279/base -> origin/gh/kwen2501/279/base 2025-12-04T08:53:58.7652885Z * [new branch] gh/kwen2501/279/head -> origin/gh/kwen2501/279/head 2025-12-04T08:53:58.7652956Z * [new branch] gh/kwen2501/279/orig -> origin/gh/kwen2501/279/orig 2025-12-04T08:53:58.7653024Z * [new branch] gh/kwen2501/280/base -> origin/gh/kwen2501/280/base 2025-12-04T08:53:58.7653099Z * [new branch] gh/kwen2501/280/head -> origin/gh/kwen2501/280/head 2025-12-04T08:53:58.7653197Z * [new branch] gh/kwen2501/280/orig -> origin/gh/kwen2501/280/orig 2025-12-04T08:53:58.7653271Z * [new branch] gh/kwen2501/281/base -> origin/gh/kwen2501/281/base 2025-12-04T08:53:58.7653343Z * [new branch] gh/kwen2501/281/head -> origin/gh/kwen2501/281/head 2025-12-04T08:53:58.7653413Z * [new branch] gh/kwen2501/281/orig -> origin/gh/kwen2501/281/orig 2025-12-04T08:53:58.7653487Z * [new branch] gh/kwen2501/282/base -> origin/gh/kwen2501/282/base 2025-12-04T08:53:58.7653559Z * [new branch] gh/kwen2501/282/head -> origin/gh/kwen2501/282/head 2025-12-04T08:53:58.7653629Z * [new branch] gh/kwen2501/282/orig -> origin/gh/kwen2501/282/orig 2025-12-04T08:53:58.7653703Z * [new branch] gh/kwen2501/283/base -> origin/gh/kwen2501/283/base 2025-12-04T08:53:58.7653774Z * [new branch] gh/kwen2501/283/head -> origin/gh/kwen2501/283/head 2025-12-04T08:53:58.7653847Z * [new branch] gh/kwen2501/283/orig -> origin/gh/kwen2501/283/orig 2025-12-04T08:53:58.7653922Z * [new branch] gh/kwen2501/284/base -> origin/gh/kwen2501/284/base 2025-12-04T08:53:58.7653993Z * [new branch] gh/kwen2501/284/head -> origin/gh/kwen2501/284/head 2025-12-04T08:53:58.7654064Z * [new branch] gh/kwen2501/284/orig -> origin/gh/kwen2501/284/orig 2025-12-04T08:53:58.7654139Z * [new branch] gh/kwen2501/285/base -> origin/gh/kwen2501/285/base 2025-12-04T08:53:58.7654211Z * [new branch] gh/kwen2501/285/head -> origin/gh/kwen2501/285/head 2025-12-04T08:53:58.7654283Z * [new branch] gh/kwen2501/285/orig -> origin/gh/kwen2501/285/orig 2025-12-04T08:53:58.7654357Z * [new branch] gh/kwen2501/286/base -> origin/gh/kwen2501/286/base 2025-12-04T08:53:58.7654427Z * [new branch] gh/kwen2501/286/head -> origin/gh/kwen2501/286/head 2025-12-04T08:53:58.7654500Z * [new branch] gh/kwen2501/286/orig -> origin/gh/kwen2501/286/orig 2025-12-04T08:53:58.7654573Z * [new branch] gh/kwen2501/287/base -> origin/gh/kwen2501/287/base 2025-12-04T08:53:58.7654644Z * [new branch] gh/kwen2501/287/head -> origin/gh/kwen2501/287/head 2025-12-04T08:53:58.7654717Z * [new branch] gh/kwen2501/287/orig -> origin/gh/kwen2501/287/orig 2025-12-04T08:53:58.7654788Z * [new branch] gh/kwen2501/288/base -> origin/gh/kwen2501/288/base 2025-12-04T08:53:58.7654888Z * [new branch] gh/kwen2501/288/head -> origin/gh/kwen2501/288/head 2025-12-04T08:53:58.7654961Z * [new branch] gh/kwen2501/288/orig -> origin/gh/kwen2501/288/orig 2025-12-04T08:53:58.7655039Z * [new branch] gh/laithsakka/251/base -> origin/gh/laithsakka/251/base 2025-12-04T08:53:58.7655117Z * [new branch] gh/laithsakka/251/head -> origin/gh/laithsakka/251/head 2025-12-04T08:53:58.7655198Z * [new branch] gh/laithsakka/251/orig -> origin/gh/laithsakka/251/orig 2025-12-04T08:53:58.7655275Z * [new branch] gh/laithsakka/276/base -> origin/gh/laithsakka/276/base 2025-12-04T08:53:58.7655350Z * [new branch] gh/laithsakka/276/head -> origin/gh/laithsakka/276/head 2025-12-04T08:53:58.7655427Z * [new branch] gh/laithsakka/276/orig -> origin/gh/laithsakka/276/orig 2025-12-04T08:53:58.7655502Z * [new branch] gh/laithsakka/28/base -> origin/gh/laithsakka/28/base 2025-12-04T08:53:58.7655579Z * [new branch] gh/laithsakka/29/base -> origin/gh/laithsakka/29/base 2025-12-04T08:53:58.7655655Z * [new branch] gh/laithsakka/30/base -> origin/gh/laithsakka/30/base 2025-12-04T08:53:58.7655731Z * [new branch] gh/laithsakka/30/head -> origin/gh/laithsakka/30/head 2025-12-04T08:53:58.7655832Z * [new branch] gh/laithsakka/31/base -> origin/gh/laithsakka/31/base 2025-12-04T08:53:58.7655910Z * [new branch] gh/laithsakka/31/head -> origin/gh/laithsakka/31/head 2025-12-04T08:53:58.7655986Z * [new branch] gh/laithsakka/313/base -> origin/gh/laithsakka/313/base 2025-12-04T08:53:58.7656062Z * [new branch] gh/laithsakka/313/head -> origin/gh/laithsakka/313/head 2025-12-04T08:53:58.7656142Z * [new branch] gh/laithsakka/313/orig -> origin/gh/laithsakka/313/orig 2025-12-04T08:53:58.7656218Z * [new branch] gh/laithsakka/316/base -> origin/gh/laithsakka/316/base 2025-12-04T08:53:58.7656297Z * [new branch] gh/laithsakka/316/head -> origin/gh/laithsakka/316/head 2025-12-04T08:53:58.7656373Z * [new branch] gh/laithsakka/316/orig -> origin/gh/laithsakka/316/orig 2025-12-04T08:53:58.7656447Z * [new branch] gh/laithsakka/317/base -> origin/gh/laithsakka/317/base 2025-12-04T08:53:58.7656526Z * [new branch] gh/laithsakka/317/head -> origin/gh/laithsakka/317/head 2025-12-04T08:53:58.7656601Z * [new branch] gh/laithsakka/317/orig -> origin/gh/laithsakka/317/orig 2025-12-04T08:53:58.7656676Z * [new branch] gh/laithsakka/319/base -> origin/gh/laithsakka/319/base 2025-12-04T08:53:58.7656807Z * [new branch] gh/laithsakka/319/head -> origin/gh/laithsakka/319/head 2025-12-04T08:53:58.7656884Z * [new branch] gh/laithsakka/319/orig -> origin/gh/laithsakka/319/orig 2025-12-04T08:53:58.7656961Z * [new branch] gh/laithsakka/32/base -> origin/gh/laithsakka/32/base 2025-12-04T08:53:58.7657040Z * [new branch] gh/laithsakka/32/head -> origin/gh/laithsakka/32/head 2025-12-04T08:53:58.7657116Z * [new branch] gh/laithsakka/320/base -> origin/gh/laithsakka/320/base 2025-12-04T08:53:58.7657191Z * [new branch] gh/laithsakka/320/head -> origin/gh/laithsakka/320/head 2025-12-04T08:53:58.7657272Z * [new branch] gh/laithsakka/320/orig -> origin/gh/laithsakka/320/orig 2025-12-04T08:53:58.7657347Z * [new branch] gh/laithsakka/321/base -> origin/gh/laithsakka/321/base 2025-12-04T08:53:58.7657421Z * [new branch] gh/laithsakka/321/head -> origin/gh/laithsakka/321/head 2025-12-04T08:53:58.7657501Z * [new branch] gh/laithsakka/321/orig -> origin/gh/laithsakka/321/orig 2025-12-04T08:53:58.7657576Z * [new branch] gh/laithsakka/322/base -> origin/gh/laithsakka/322/base 2025-12-04T08:53:58.7657687Z * [new branch] gh/laithsakka/322/head -> origin/gh/laithsakka/322/head 2025-12-04T08:53:58.7657764Z * [new branch] gh/laithsakka/322/orig -> origin/gh/laithsakka/322/orig 2025-12-04T08:53:58.7657839Z * [new branch] gh/laithsakka/323/base -> origin/gh/laithsakka/323/base 2025-12-04T08:53:58.7657914Z * [new branch] gh/laithsakka/323/head -> origin/gh/laithsakka/323/head 2025-12-04T08:53:58.7657993Z * [new branch] gh/laithsakka/323/orig -> origin/gh/laithsakka/323/orig 2025-12-04T08:53:58.7658068Z * [new branch] gh/laithsakka/324/base -> origin/gh/laithsakka/324/base 2025-12-04T08:53:58.7658146Z * [new branch] gh/laithsakka/324/head -> origin/gh/laithsakka/324/head 2025-12-04T08:53:58.7658221Z * [new branch] gh/laithsakka/324/orig -> origin/gh/laithsakka/324/orig 2025-12-04T08:53:58.7658296Z * [new branch] gh/laithsakka/325/base -> origin/gh/laithsakka/325/base 2025-12-04T08:53:58.7658376Z * [new branch] gh/laithsakka/325/head -> origin/gh/laithsakka/325/head 2025-12-04T08:53:58.7658450Z * [new branch] gh/laithsakka/325/orig -> origin/gh/laithsakka/325/orig 2025-12-04T08:53:58.7658525Z * [new branch] gh/laithsakka/326/base -> origin/gh/laithsakka/326/base 2025-12-04T08:53:58.7658644Z * [new branch] gh/laithsakka/326/head -> origin/gh/laithsakka/326/head 2025-12-04T08:53:58.7658720Z * [new branch] gh/laithsakka/326/orig -> origin/gh/laithsakka/326/orig 2025-12-04T08:53:58.7658795Z * [new branch] gh/laithsakka/327/base -> origin/gh/laithsakka/327/base 2025-12-04T08:53:58.7658872Z * [new branch] gh/laithsakka/327/head -> origin/gh/laithsakka/327/head 2025-12-04T08:53:58.7658945Z * [new branch] gh/laithsakka/327/orig -> origin/gh/laithsakka/327/orig 2025-12-04T08:53:58.7659020Z * [new branch] gh/laithsakka/328/base -> origin/gh/laithsakka/328/base 2025-12-04T08:53:58.7659102Z * [new branch] gh/laithsakka/328/head -> origin/gh/laithsakka/328/head 2025-12-04T08:53:58.7659180Z * [new branch] gh/laithsakka/328/orig -> origin/gh/laithsakka/328/orig 2025-12-04T08:53:58.7659253Z * [new branch] gh/liangel/4/base -> origin/gh/liangel/4/base 2025-12-04T08:53:58.7659328Z * [new branch] gh/liangel/4/head -> origin/gh/liangel/4/head 2025-12-04T08:53:58.7659399Z * [new branch] gh/liangel/4/orig -> origin/gh/liangel/4/orig 2025-12-04T08:53:58.7659476Z * [new branch] gh/lucaskabela/1/base -> origin/gh/lucaskabela/1/base 2025-12-04T08:53:58.7659553Z * [new branch] gh/lucaskabela/1/head -> origin/gh/lucaskabela/1/head 2025-12-04T08:53:58.7659619Z * [new branch] gh/lw/4/base -> origin/gh/lw/4/base 2025-12-04T08:53:58.7659688Z * [new branch] gh/lw/4/head -> origin/gh/lw/4/head 2025-12-04T08:53:58.7659755Z * [new branch] gh/lw/4/orig -> origin/gh/lw/4/orig 2025-12-04T08:53:58.7659819Z * [new branch] gh/lw/5/base -> origin/gh/lw/5/base 2025-12-04T08:53:58.7659884Z * [new branch] gh/lw/5/head -> origin/gh/lw/5/head 2025-12-04T08:53:58.7659948Z * [new branch] gh/lw/5/orig -> origin/gh/lw/5/orig 2025-12-04T08:53:58.7660012Z * [new branch] gh/lw/6/base -> origin/gh/lw/6/base 2025-12-04T08:53:58.7660077Z * [new branch] gh/lw/6/head -> origin/gh/lw/6/head 2025-12-04T08:53:58.7660140Z * [new branch] gh/lw/6/orig -> origin/gh/lw/6/orig 2025-12-04T08:53:58.7660211Z * [new branch] gh/malfet/14/base -> origin/gh/malfet/14/base 2025-12-04T08:53:58.7660286Z * [new branch] gh/malfet/417/base -> origin/gh/malfet/417/base 2025-12-04T08:53:58.7660387Z * [new branch] gh/malfet/417/head -> origin/gh/malfet/417/head 2025-12-04T08:53:58.7660460Z * [new branch] gh/malfet/417/orig -> origin/gh/malfet/417/orig 2025-12-04T08:53:58.7660534Z * [new branch] gh/malfet/506/base -> origin/gh/malfet/506/base 2025-12-04T08:53:58.7660603Z * [new branch] gh/malfet/506/head -> origin/gh/malfet/506/head 2025-12-04T08:53:58.7660677Z * [new branch] gh/malfet/506/orig -> origin/gh/malfet/506/orig 2025-12-04T08:53:58.7660747Z * [new branch] gh/malfet/517/base -> origin/gh/malfet/517/base 2025-12-04T08:53:58.7660816Z * [new branch] gh/malfet/517/head -> origin/gh/malfet/517/head 2025-12-04T08:53:58.7660885Z * [new branch] gh/malfet/528/base -> origin/gh/malfet/528/base 2025-12-04T08:53:58.7660958Z * [new branch] gh/malfet/528/head -> origin/gh/malfet/528/head 2025-12-04T08:53:58.7661029Z * [new branch] gh/malfet/528/orig -> origin/gh/malfet/528/orig 2025-12-04T08:53:58.7661097Z * [new branch] gh/malfet/537/base -> origin/gh/malfet/537/base 2025-12-04T08:53:58.7661169Z * [new branch] gh/malfet/537/head -> origin/gh/malfet/537/head 2025-12-04T08:53:58.7661240Z * [new branch] gh/malfet/537/orig -> origin/gh/malfet/537/orig 2025-12-04T08:53:58.7661338Z * [new branch] gh/malfet/546/base -> origin/gh/malfet/546/base 2025-12-04T08:53:58.7661410Z * [new branch] gh/malfet/546/head -> origin/gh/malfet/546/head 2025-12-04T08:53:58.7661481Z * [new branch] gh/malfet/546/orig -> origin/gh/malfet/546/orig 2025-12-04T08:53:58.7661552Z * [new branch] gh/malfet/565/base -> origin/gh/malfet/565/base 2025-12-04T08:53:58.7661621Z * [new branch] gh/malfet/565/head -> origin/gh/malfet/565/head 2025-12-04T08:53:58.7661694Z * [new branch] gh/malfet/565/orig -> origin/gh/malfet/565/orig 2025-12-04T08:53:58.7661767Z * [new branch] gh/malfet/575/base -> origin/gh/malfet/575/base 2025-12-04T08:53:58.7661836Z * [new branch] gh/malfet/575/head -> origin/gh/malfet/575/head 2025-12-04T08:53:58.7661906Z * [new branch] gh/malfet/575/orig -> origin/gh/malfet/575/orig 2025-12-04T08:53:58.7661979Z * [new branch] gh/malfet/580/base -> origin/gh/malfet/580/base 2025-12-04T08:53:58.7662050Z * [new branch] gh/malfet/580/head -> origin/gh/malfet/580/head 2025-12-04T08:53:58.7662118Z * [new branch] gh/malfet/580/orig -> origin/gh/malfet/580/orig 2025-12-04T08:53:58.7662189Z * [new branch] gh/malfet/581/base -> origin/gh/malfet/581/base 2025-12-04T08:53:58.7662259Z * [new branch] gh/malfet/581/head -> origin/gh/malfet/581/head 2025-12-04T08:53:58.7662330Z * [new branch] gh/malfet/581/orig -> origin/gh/malfet/581/orig 2025-12-04T08:53:58.7662400Z * [new branch] gh/malfet/583/base -> origin/gh/malfet/583/base 2025-12-04T08:53:58.7662470Z * [new branch] gh/malfet/583/head -> origin/gh/malfet/583/head 2025-12-04T08:53:58.7662540Z * [new branch] gh/malfet/583/orig -> origin/gh/malfet/583/orig 2025-12-04T08:53:58.7662613Z * [new branch] gh/malfet/586/base -> origin/gh/malfet/586/base 2025-12-04T08:53:58.7662683Z * [new branch] gh/malfet/586/head -> origin/gh/malfet/586/head 2025-12-04T08:53:58.7662751Z * [new branch] gh/malfet/586/orig -> origin/gh/malfet/586/orig 2025-12-04T08:53:58.7662824Z * [new branch] gh/malfet/587/base -> origin/gh/malfet/587/base 2025-12-04T08:53:58.7662892Z * [new branch] gh/malfet/587/head -> origin/gh/malfet/587/head 2025-12-04T08:53:58.7662994Z * [new branch] gh/malfet/587/orig -> origin/gh/malfet/587/orig 2025-12-04T08:53:58.7663065Z * [new branch] gh/malfet/588/base -> origin/gh/malfet/588/base 2025-12-04T08:53:58.7663133Z * [new branch] gh/malfet/588/head -> origin/gh/malfet/588/head 2025-12-04T08:53:58.7663207Z * [new branch] gh/malfet/588/orig -> origin/gh/malfet/588/orig 2025-12-04T08:53:58.7663278Z * [new branch] gh/malfet/589/base -> origin/gh/malfet/589/base 2025-12-04T08:53:58.7663347Z * [new branch] gh/malfet/589/head -> origin/gh/malfet/589/head 2025-12-04T08:53:58.7663419Z * [new branch] gh/malfet/589/orig -> origin/gh/malfet/589/orig 2025-12-04T08:53:58.7663488Z * [new branch] gh/malfet/590/base -> origin/gh/malfet/590/base 2025-12-04T08:53:58.7663557Z * [new branch] gh/malfet/590/head -> origin/gh/malfet/590/head 2025-12-04T08:53:58.7663627Z * [new branch] gh/malfet/590/orig -> origin/gh/malfet/590/orig 2025-12-04T08:53:58.7663699Z * [new branch] gh/malfet/591/base -> origin/gh/malfet/591/base 2025-12-04T08:53:58.7663768Z * [new branch] gh/malfet/591/head -> origin/gh/malfet/591/head 2025-12-04T08:53:58.7663839Z * [new branch] gh/malfet/591/orig -> origin/gh/malfet/591/orig 2025-12-04T08:53:58.7663933Z * [new branch] gh/malfet/592/base -> origin/gh/malfet/592/base 2025-12-04T08:53:58.7664003Z * [new branch] gh/malfet/592/head -> origin/gh/malfet/592/head 2025-12-04T08:53:58.7664078Z * [new branch] gh/malfet/592/orig -> origin/gh/malfet/592/orig 2025-12-04T08:53:58.7664146Z * [new branch] gh/malfet/593/base -> origin/gh/malfet/593/base 2025-12-04T08:53:58.7664216Z * [new branch] gh/malfet/593/head -> origin/gh/malfet/593/head 2025-12-04T08:53:58.7664288Z * [new branch] gh/malfet/593/orig -> origin/gh/malfet/593/orig 2025-12-04T08:53:58.7664359Z * [new branch] gh/malfet/594/base -> origin/gh/malfet/594/base 2025-12-04T08:53:58.7664429Z * [new branch] gh/malfet/594/head -> origin/gh/malfet/594/head 2025-12-04T08:53:58.7664501Z * [new branch] gh/malfet/594/orig -> origin/gh/malfet/594/orig 2025-12-04T08:53:58.7664572Z * [new branch] gh/malfet/595/base -> origin/gh/malfet/595/base 2025-12-04T08:53:58.7664640Z * [new branch] gh/malfet/595/head -> origin/gh/malfet/595/head 2025-12-04T08:53:58.7664712Z * [new branch] gh/malfet/595/orig -> origin/gh/malfet/595/orig 2025-12-04T08:53:58.7664781Z * [new branch] gh/malfet/596/base -> origin/gh/malfet/596/base 2025-12-04T08:53:58.7664852Z * [new branch] gh/malfet/596/head -> origin/gh/malfet/596/head 2025-12-04T08:53:58.7664921Z * [new branch] gh/malfet/596/orig -> origin/gh/malfet/596/orig 2025-12-04T08:53:58.7664991Z * [new branch] gh/malfet/597/base -> origin/gh/malfet/597/base 2025-12-04T08:53:58.7665062Z * [new branch] gh/malfet/597/head -> origin/gh/malfet/597/head 2025-12-04T08:53:58.7665131Z * [new branch] gh/malfet/597/orig -> origin/gh/malfet/597/orig 2025-12-04T08:53:58.7665201Z * [new branch] gh/malfet/598/base -> origin/gh/malfet/598/base 2025-12-04T08:53:58.7665273Z * [new branch] gh/malfet/598/head -> origin/gh/malfet/598/head 2025-12-04T08:53:58.7665341Z * [new branch] gh/malfet/598/orig -> origin/gh/malfet/598/orig 2025-12-04T08:53:58.7665410Z * [new branch] gh/malfet/599/base -> origin/gh/malfet/599/base 2025-12-04T08:53:58.7665483Z * [new branch] gh/malfet/599/head -> origin/gh/malfet/599/head 2025-12-04T08:53:58.7665553Z * [new branch] gh/malfet/599/orig -> origin/gh/malfet/599/orig 2025-12-04T08:53:58.7665648Z * [new branch] gh/malfet/600/base -> origin/gh/malfet/600/base 2025-12-04T08:53:58.7665720Z * [new branch] gh/malfet/600/head -> origin/gh/malfet/600/head 2025-12-04T08:53:58.7665790Z * [new branch] gh/malfet/600/orig -> origin/gh/malfet/600/orig 2025-12-04T08:53:58.7665860Z * [new branch] gh/malfet/601/base -> origin/gh/malfet/601/base 2025-12-04T08:53:58.7665929Z * [new branch] gh/malfet/601/head -> origin/gh/malfet/601/head 2025-12-04T08:53:58.7665998Z * [new branch] gh/malfet/601/orig -> origin/gh/malfet/601/orig 2025-12-04T08:53:58.7666067Z * [new branch] gh/malfet/602/base -> origin/gh/malfet/602/base 2025-12-04T08:53:58.7666138Z * [new branch] gh/malfet/602/head -> origin/gh/malfet/602/head 2025-12-04T08:53:58.7666207Z * [new branch] gh/malfet/602/orig -> origin/gh/malfet/602/orig 2025-12-04T08:53:58.7666277Z * [new branch] gh/malfet/603/base -> origin/gh/malfet/603/base 2025-12-04T08:53:58.7666349Z * [new branch] gh/malfet/603/head -> origin/gh/malfet/603/head 2025-12-04T08:53:58.7666418Z * [new branch] gh/malfet/603/orig -> origin/gh/malfet/603/orig 2025-12-04T08:53:58.7666514Z * [new branch] gh/malfet/604/base -> origin/gh/malfet/604/base 2025-12-04T08:53:58.7666585Z * [new branch] gh/malfet/604/head -> origin/gh/malfet/604/head 2025-12-04T08:53:58.7666652Z * [new branch] gh/malfet/604/orig -> origin/gh/malfet/604/orig 2025-12-04T08:53:58.7666724Z * [new branch] gh/malfet/605/base -> origin/gh/malfet/605/base 2025-12-04T08:53:58.7666829Z * [new branch] gh/malfet/605/head -> origin/gh/malfet/605/head 2025-12-04T08:53:58.7666900Z * [new branch] gh/malfet/605/orig -> origin/gh/malfet/605/orig 2025-12-04T08:53:58.7666973Z * [new branch] gh/malfet/606/base -> origin/gh/malfet/606/base 2025-12-04T08:53:58.7667042Z * [new branch] gh/malfet/606/head -> origin/gh/malfet/606/head 2025-12-04T08:53:58.7667111Z * [new branch] gh/malfet/606/orig -> origin/gh/malfet/606/orig 2025-12-04T08:53:58.7667185Z * [new branch] gh/malfet/607/base -> origin/gh/malfet/607/base 2025-12-04T08:53:58.7667254Z * [new branch] gh/malfet/607/head -> origin/gh/malfet/607/head 2025-12-04T08:53:58.7667324Z * [new branch] gh/malfet/607/orig -> origin/gh/malfet/607/orig 2025-12-04T08:53:58.7667396Z * [new branch] gh/malfet/608/base -> origin/gh/malfet/608/base 2025-12-04T08:53:58.7667470Z * [new branch] gh/malfet/608/head -> origin/gh/malfet/608/head 2025-12-04T08:53:58.7667540Z * [new branch] gh/malfet/608/orig -> origin/gh/malfet/608/orig 2025-12-04T08:53:58.7667614Z * [new branch] gh/malfet/609/base -> origin/gh/malfet/609/base 2025-12-04T08:53:58.7667772Z * [new branch] gh/malfet/609/head -> origin/gh/malfet/609/head 2025-12-04T08:53:58.7671637Z * [new branch] gh/malfet/609/orig -> origin/gh/malfet/609/orig 2025-12-04T08:53:58.7671727Z * [new branch] gh/malfet/610/base -> origin/gh/malfet/610/base 2025-12-04T08:53:58.7671801Z * [new branch] gh/malfet/610/head -> origin/gh/malfet/610/head 2025-12-04T08:53:58.7671869Z * [new branch] gh/malfet/610/orig -> origin/gh/malfet/610/orig 2025-12-04T08:53:58.7671941Z * [new branch] gh/malfet/611/base -> origin/gh/malfet/611/base 2025-12-04T08:53:58.7672009Z * [new branch] gh/malfet/611/head -> origin/gh/malfet/611/head 2025-12-04T08:53:58.7672078Z * [new branch] gh/malfet/611/orig -> origin/gh/malfet/611/orig 2025-12-04T08:53:58.7672213Z * [new branch] gh/malfet/612/base -> origin/gh/malfet/612/base 2025-12-04T08:53:58.7672282Z * [new branch] gh/malfet/612/head -> origin/gh/malfet/612/head 2025-12-04T08:53:58.7672351Z * [new branch] gh/malfet/612/orig -> origin/gh/malfet/612/orig 2025-12-04T08:53:58.7672425Z * [new branch] gh/malfet/64/base -> origin/gh/malfet/64/base 2025-12-04T08:53:58.7672494Z * [new branch] gh/malfet/64/head -> origin/gh/malfet/64/head 2025-12-04T08:53:58.7672586Z * [new branch] gh/manuelcandales/11/base -> origin/gh/manuelcandales/11/base 2025-12-04T08:53:58.7672673Z * [new branch] gh/manuelcandales/11/head -> origin/gh/manuelcandales/11/head 2025-12-04T08:53:58.7672756Z * [new branch] gh/manuelcandales/11/orig -> origin/gh/manuelcandales/11/orig 2025-12-04T08:53:58.7672832Z * [new branch] gh/markkm/1/base -> origin/gh/markkm/1/base 2025-12-04T08:53:58.7672909Z * [new branch] gh/masnesral/1/base -> origin/gh/masnesral/1/base 2025-12-04T08:53:58.7672983Z * [new branch] gh/masnesral/1/head -> origin/gh/masnesral/1/head 2025-12-04T08:53:58.7673056Z * [new branch] gh/masnesral/1/orig -> origin/gh/masnesral/1/orig 2025-12-04T08:53:58.7673194Z * [new branch] gh/mhorowitz/0/base -> origin/gh/mhorowitz/0/base 2025-12-04T08:53:58.7673264Z * [new branch] gh/mhorowitz/0/head -> origin/gh/mhorowitz/0/head 2025-12-04T08:53:58.7673337Z * [new branch] gh/mhorowitz/1/base -> origin/gh/mhorowitz/1/base 2025-12-04T08:53:58.7673407Z * [new branch] gh/mhorowitz/1/head -> origin/gh/mhorowitz/1/head 2025-12-04T08:53:58.7673476Z * [new branch] gh/mhorowitz/2/base -> origin/gh/mhorowitz/2/base 2025-12-04T08:53:58.7673548Z * [new branch] gh/mhorowitz/2/head -> origin/gh/mhorowitz/2/head 2025-12-04T08:53:58.7673621Z * [new branch] gh/mhorowitz/3/base -> origin/gh/mhorowitz/3/base 2025-12-04T08:53:58.7673691Z * [new branch] gh/mhorowitz/3/head -> origin/gh/mhorowitz/3/head 2025-12-04T08:53:58.7673762Z * [new branch] gh/mhorowitz/4/base -> origin/gh/mhorowitz/4/base 2025-12-04T08:53:58.7673834Z * [new branch] gh/mhorowitz/4/head -> origin/gh/mhorowitz/4/head 2025-12-04T08:53:58.7673905Z * [new branch] gh/mhorowitz/5/base -> origin/gh/mhorowitz/5/base 2025-12-04T08:53:58.7673976Z * [new branch] gh/mhorowitz/5/head -> origin/gh/mhorowitz/5/head 2025-12-04T08:53:58.7674046Z * [new branch] gh/mhorowitz/6/base -> origin/gh/mhorowitz/6/base 2025-12-04T08:53:58.7674116Z * [new branch] gh/mhorowitz/6/head -> origin/gh/mhorowitz/6/head 2025-12-04T08:53:58.7674221Z * [new branch] gh/mikaylagawarecki/234/base -> origin/gh/mikaylagawarecki/234/base 2025-12-04T08:53:58.7674321Z * [new branch] gh/mikaylagawarecki/234/head -> origin/gh/mikaylagawarecki/234/head 2025-12-04T08:53:58.7674418Z * [new branch] gh/mikaylagawarecki/235/base -> origin/gh/mikaylagawarecki/235/base 2025-12-04T08:53:58.7674514Z * [new branch] gh/mikaylagawarecki/235/head -> origin/gh/mikaylagawarecki/235/head 2025-12-04T08:53:58.7674608Z * [new branch] gh/mikaylagawarecki/236/base -> origin/gh/mikaylagawarecki/236/base 2025-12-04T08:53:58.7674703Z * [new branch] gh/mikaylagawarecki/236/head -> origin/gh/mikaylagawarecki/236/head 2025-12-04T08:53:58.7674795Z * [new branch] gh/mikaylagawarecki/237/base -> origin/gh/mikaylagawarecki/237/base 2025-12-04T08:53:58.7674888Z * [new branch] gh/mikaylagawarecki/237/head -> origin/gh/mikaylagawarecki/237/head 2025-12-04T08:53:58.7674982Z * [new branch] gh/mikaylagawarecki/238/base -> origin/gh/mikaylagawarecki/238/base 2025-12-04T08:53:58.7675098Z * [new branch] gh/mikaylagawarecki/238/head -> origin/gh/mikaylagawarecki/238/head 2025-12-04T08:53:58.7675191Z * [new branch] gh/mikaylagawarecki/336/base -> origin/gh/mikaylagawarecki/336/base 2025-12-04T08:53:58.7675286Z * [new branch] gh/mikaylagawarecki/336/head -> origin/gh/mikaylagawarecki/336/head 2025-12-04T08:53:58.7675381Z * [new branch] gh/mikaylagawarecki/336/orig -> origin/gh/mikaylagawarecki/336/orig 2025-12-04T08:53:58.7675473Z * [new branch] gh/mikaylagawarecki/341/base -> origin/gh/mikaylagawarecki/341/base 2025-12-04T08:53:58.7675566Z * [new branch] gh/mikaylagawarecki/341/head -> origin/gh/mikaylagawarecki/341/head 2025-12-04T08:53:58.7675659Z * [new branch] gh/mikaylagawarecki/341/orig -> origin/gh/mikaylagawarecki/341/orig 2025-12-04T08:53:58.7675752Z * [new branch] gh/mikaylagawarecki/342/base -> origin/gh/mikaylagawarecki/342/base 2025-12-04T08:53:58.7675847Z * [new branch] gh/mikaylagawarecki/342/head -> origin/gh/mikaylagawarecki/342/head 2025-12-04T08:53:58.7675939Z * [new branch] gh/mikaylagawarecki/342/orig -> origin/gh/mikaylagawarecki/342/orig 2025-12-04T08:53:58.7676031Z * [new branch] gh/mikaylagawarecki/345/base -> origin/gh/mikaylagawarecki/345/base 2025-12-04T08:53:58.7676152Z * [new branch] gh/mikaylagawarecki/345/head -> origin/gh/mikaylagawarecki/345/head 2025-12-04T08:53:58.7676245Z * [new branch] gh/mikaylagawarecki/345/orig -> origin/gh/mikaylagawarecki/345/orig 2025-12-04T08:53:58.7676337Z * [new branch] gh/mikaylagawarecki/346/base -> origin/gh/mikaylagawarecki/346/base 2025-12-04T08:53:58.7676431Z * [new branch] gh/mikaylagawarecki/346/head -> origin/gh/mikaylagawarecki/346/head 2025-12-04T08:53:58.7676524Z * [new branch] gh/mikaylagawarecki/346/orig -> origin/gh/mikaylagawarecki/346/orig 2025-12-04T08:53:58.7676622Z * [new branch] gh/mikaylagawarecki/347/base -> origin/gh/mikaylagawarecki/347/base 2025-12-04T08:53:58.7676714Z * [new branch] gh/mikaylagawarecki/347/head -> origin/gh/mikaylagawarecki/347/head 2025-12-04T08:53:58.7676858Z * [new branch] gh/mikaylagawarecki/347/orig -> origin/gh/mikaylagawarecki/347/orig 2025-12-04T08:53:58.7676955Z * [new branch] gh/mikaylagawarecki/350/base -> origin/gh/mikaylagawarecki/350/base 2025-12-04T08:53:58.7677047Z * [new branch] gh/mikaylagawarecki/350/head -> origin/gh/mikaylagawarecki/350/head 2025-12-04T08:53:58.7677138Z * [new branch] gh/mikaylagawarecki/350/orig -> origin/gh/mikaylagawarecki/350/orig 2025-12-04T08:53:58.7677232Z * [new branch] gh/mikaylagawarecki/351/base -> origin/gh/mikaylagawarecki/351/base 2025-12-04T08:53:58.7677324Z * [new branch] gh/mikaylagawarecki/351/head -> origin/gh/mikaylagawarecki/351/head 2025-12-04T08:53:58.7677417Z * [new branch] gh/mikaylagawarecki/351/orig -> origin/gh/mikaylagawarecki/351/orig 2025-12-04T08:53:58.7677510Z * [new branch] gh/mikaylagawarecki/352/base -> origin/gh/mikaylagawarecki/352/base 2025-12-04T08:53:58.7677603Z * [new branch] gh/mikaylagawarecki/352/head -> origin/gh/mikaylagawarecki/352/head 2025-12-04T08:53:58.7677695Z * [new branch] gh/mikaylagawarecki/352/orig -> origin/gh/mikaylagawarecki/352/orig 2025-12-04T08:53:58.7677789Z * [new branch] gh/mikaylagawarecki/353/base -> origin/gh/mikaylagawarecki/353/base 2025-12-04T08:53:58.7677881Z * [new branch] gh/mikaylagawarecki/353/head -> origin/gh/mikaylagawarecki/353/head 2025-12-04T08:53:58.7677974Z * [new branch] gh/mikaylagawarecki/353/orig -> origin/gh/mikaylagawarecki/353/orig 2025-12-04T08:53:58.7678068Z * [new branch] gh/mikaylagawarecki/354/base -> origin/gh/mikaylagawarecki/354/base 2025-12-04T08:53:58.7678201Z * [new branch] gh/mikaylagawarecki/354/head -> origin/gh/mikaylagawarecki/354/head 2025-12-04T08:53:58.7678294Z * [new branch] gh/mikaylagawarecki/354/orig -> origin/gh/mikaylagawarecki/354/orig 2025-12-04T08:53:58.7678387Z * [new branch] gh/mikaylagawarecki/356/base -> origin/gh/mikaylagawarecki/356/base 2025-12-04T08:53:58.7678480Z * [new branch] gh/mikaylagawarecki/356/head -> origin/gh/mikaylagawarecki/356/head 2025-12-04T08:53:58.7678574Z * [new branch] gh/mikaylagawarecki/356/orig -> origin/gh/mikaylagawarecki/356/orig 2025-12-04T08:53:58.7678665Z * [new branch] gh/mikaylagawarecki/357/base -> origin/gh/mikaylagawarecki/357/base 2025-12-04T08:53:58.7678757Z * [new branch] gh/mikaylagawarecki/357/head -> origin/gh/mikaylagawarecki/357/head 2025-12-04T08:53:58.7678850Z * [new branch] gh/mikaylagawarecki/357/orig -> origin/gh/mikaylagawarecki/357/orig 2025-12-04T08:53:58.7678943Z * [new branch] gh/mikaylagawarecki/359/base -> origin/gh/mikaylagawarecki/359/base 2025-12-04T08:53:58.7679038Z * [new branch] gh/mikaylagawarecki/359/head -> origin/gh/mikaylagawarecki/359/head 2025-12-04T08:53:58.7679133Z * [new branch] gh/mikaylagawarecki/359/orig -> origin/gh/mikaylagawarecki/359/orig 2025-12-04T08:53:58.7679266Z * [new branch] gh/mikaylagawarecki/360/base -> origin/gh/mikaylagawarecki/360/base 2025-12-04T08:53:58.7679359Z * [new branch] gh/mikaylagawarecki/360/head -> origin/gh/mikaylagawarecki/360/head 2025-12-04T08:53:58.7679456Z * [new branch] gh/mikaylagawarecki/360/orig -> origin/gh/mikaylagawarecki/360/orig 2025-12-04T08:53:58.7679550Z * [new branch] gh/mikaylagawarecki/361/base -> origin/gh/mikaylagawarecki/361/base 2025-12-04T08:53:58.7679644Z * [new branch] gh/mikaylagawarecki/361/head -> origin/gh/mikaylagawarecki/361/head 2025-12-04T08:53:58.7679739Z * [new branch] gh/mikaylagawarecki/361/orig -> origin/gh/mikaylagawarecki/361/orig 2025-12-04T08:53:58.7679833Z * [new branch] gh/mikaylagawarecki/362/base -> origin/gh/mikaylagawarecki/362/base 2025-12-04T08:53:58.7679926Z * [new branch] gh/mikaylagawarecki/362/head -> origin/gh/mikaylagawarecki/362/head 2025-12-04T08:53:58.7680019Z * [new branch] gh/mikaylagawarecki/362/orig -> origin/gh/mikaylagawarecki/362/orig 2025-12-04T08:53:58.7680112Z * [new branch] gh/mikaylagawarecki/363/base -> origin/gh/mikaylagawarecki/363/base 2025-12-04T08:53:58.7680206Z * [new branch] gh/mikaylagawarecki/363/head -> origin/gh/mikaylagawarecki/363/head 2025-12-04T08:53:58.7680298Z * [new branch] gh/mikaylagawarecki/363/orig -> origin/gh/mikaylagawarecki/363/orig 2025-12-04T08:53:58.7680390Z * [new branch] gh/mikaylagawarecki/364/base -> origin/gh/mikaylagawarecki/364/base 2025-12-04T08:53:58.7680485Z * [new branch] gh/mikaylagawarecki/364/head -> origin/gh/mikaylagawarecki/364/head 2025-12-04T08:53:58.7680578Z * [new branch] gh/mikaylagawarecki/364/orig -> origin/gh/mikaylagawarecki/364/orig 2025-12-04T08:53:58.7680670Z * [new branch] gh/mikaylagawarecki/365/base -> origin/gh/mikaylagawarecki/365/base 2025-12-04T08:53:58.7680766Z * [new branch] gh/mikaylagawarecki/365/head -> origin/gh/mikaylagawarecki/365/head 2025-12-04T08:53:58.7680858Z * [new branch] gh/mikaylagawarecki/365/orig -> origin/gh/mikaylagawarecki/365/orig 2025-12-04T08:53:58.7680950Z * [new branch] gh/mikaylagawarecki/366/base -> origin/gh/mikaylagawarecki/366/base 2025-12-04T08:53:58.7681044Z * [new branch] gh/mikaylagawarecki/366/head -> origin/gh/mikaylagawarecki/366/head 2025-12-04T08:53:58.7681136Z * [new branch] gh/mikaylagawarecki/366/orig -> origin/gh/mikaylagawarecki/366/orig 2025-12-04T08:53:58.7681254Z * [new branch] gh/mikaylagawarecki/367/base -> origin/gh/mikaylagawarecki/367/base 2025-12-04T08:53:58.7681346Z * [new branch] gh/mikaylagawarecki/367/head -> origin/gh/mikaylagawarecki/367/head 2025-12-04T08:53:58.7681438Z * [new branch] gh/mikaylagawarecki/367/orig -> origin/gh/mikaylagawarecki/367/orig 2025-12-04T08:53:58.7681533Z * [new branch] gh/mikaylagawarecki/368/base -> origin/gh/mikaylagawarecki/368/base 2025-12-04T08:53:58.7681625Z * [new branch] gh/mikaylagawarecki/368/head -> origin/gh/mikaylagawarecki/368/head 2025-12-04T08:53:58.7681717Z * [new branch] gh/mikaylagawarecki/368/orig -> origin/gh/mikaylagawarecki/368/orig 2025-12-04T08:53:58.7681811Z * [new branch] gh/mikaylagawarecki/369/base -> origin/gh/mikaylagawarecki/369/base 2025-12-04T08:53:58.7681903Z * [new branch] gh/mikaylagawarecki/369/head -> origin/gh/mikaylagawarecki/369/head 2025-12-04T08:53:58.7681998Z * [new branch] gh/mikaylagawarecki/369/orig -> origin/gh/mikaylagawarecki/369/orig 2025-12-04T08:53:58.7682095Z * [new branch] gh/mikaylagawarecki/370/base -> origin/gh/mikaylagawarecki/370/base 2025-12-04T08:53:58.7682189Z * [new branch] gh/mikaylagawarecki/370/head -> origin/gh/mikaylagawarecki/370/head 2025-12-04T08:53:58.7682312Z * [new branch] gh/mikaylagawarecki/370/orig -> origin/gh/mikaylagawarecki/370/orig 2025-12-04T08:53:58.7682408Z * [new branch] gh/mikaylagawarecki/371/base -> origin/gh/mikaylagawarecki/371/base 2025-12-04T08:53:58.7682503Z * [new branch] gh/mikaylagawarecki/371/head -> origin/gh/mikaylagawarecki/371/head 2025-12-04T08:53:58.7682599Z * [new branch] gh/mikaylagawarecki/371/orig -> origin/gh/mikaylagawarecki/371/orig 2025-12-04T08:53:58.7682693Z * [new branch] gh/mikaylagawarecki/372/base -> origin/gh/mikaylagawarecki/372/base 2025-12-04T08:53:58.7682787Z * [new branch] gh/mikaylagawarecki/372/head -> origin/gh/mikaylagawarecki/372/head 2025-12-04T08:53:58.7682880Z * [new branch] gh/mikaylagawarecki/372/orig -> origin/gh/mikaylagawarecki/372/orig 2025-12-04T08:53:58.7682974Z * [new branch] gh/mikaylagawarecki/373/base -> origin/gh/mikaylagawarecki/373/base 2025-12-04T08:53:58.7683068Z * [new branch] gh/mikaylagawarecki/373/head -> origin/gh/mikaylagawarecki/373/head 2025-12-04T08:53:58.7683161Z * [new branch] gh/mikaylagawarecki/373/orig -> origin/gh/mikaylagawarecki/373/orig 2025-12-04T08:53:58.7683254Z * [new branch] gh/mikaylagawarecki/374/base -> origin/gh/mikaylagawarecki/374/base 2025-12-04T08:53:58.7683347Z * [new branch] gh/mikaylagawarecki/374/head -> origin/gh/mikaylagawarecki/374/head 2025-12-04T08:53:58.7683440Z * [new branch] gh/mikaylagawarecki/374/orig -> origin/gh/mikaylagawarecki/374/orig 2025-12-04T08:53:58.7683533Z * [new branch] gh/mikaylagawarecki/375/base -> origin/gh/mikaylagawarecki/375/base 2025-12-04T08:53:58.7683627Z * [new branch] gh/mikaylagawarecki/375/head -> origin/gh/mikaylagawarecki/375/head 2025-12-04T08:53:58.7683720Z * [new branch] gh/mikaylagawarecki/375/orig -> origin/gh/mikaylagawarecki/375/orig 2025-12-04T08:53:58.7683815Z * [new branch] gh/mikaylagawarecki/376/base -> origin/gh/mikaylagawarecki/376/base 2025-12-04T08:53:58.7683907Z * [new branch] gh/mikaylagawarecki/376/head -> origin/gh/mikaylagawarecki/376/head 2025-12-04T08:53:58.7684001Z * [new branch] gh/mikaylagawarecki/376/orig -> origin/gh/mikaylagawarecki/376/orig 2025-12-04T08:53:58.7684093Z * [new branch] gh/mikaylagawarecki/377/base -> origin/gh/mikaylagawarecki/377/base 2025-12-04T08:53:58.7684189Z * [new branch] gh/mikaylagawarecki/377/head -> origin/gh/mikaylagawarecki/377/head 2025-12-04T08:53:58.7684309Z * [new branch] gh/mikaylagawarecki/377/orig -> origin/gh/mikaylagawarecki/377/orig 2025-12-04T08:53:58.7684401Z * [new branch] gh/mikaylagawarecki/378/base -> origin/gh/mikaylagawarecki/378/base 2025-12-04T08:53:58.7684494Z * [new branch] gh/mikaylagawarecki/378/head -> origin/gh/mikaylagawarecki/378/head 2025-12-04T08:53:58.7684590Z * [new branch] gh/mikaylagawarecki/378/orig -> origin/gh/mikaylagawarecki/378/orig 2025-12-04T08:53:58.7684683Z * [new branch] gh/mikaylagawarecki/379/base -> origin/gh/mikaylagawarecki/379/base 2025-12-04T08:53:58.7684776Z * [new branch] gh/mikaylagawarecki/379/head -> origin/gh/mikaylagawarecki/379/head 2025-12-04T08:53:58.7684869Z * [new branch] gh/mikaylagawarecki/379/orig -> origin/gh/mikaylagawarecki/379/orig 2025-12-04T08:53:58.7684961Z * [new branch] gh/mikaylagawarecki/380/base -> origin/gh/mikaylagawarecki/380/base 2025-12-04T08:53:58.7685058Z * [new branch] gh/mikaylagawarecki/380/head -> origin/gh/mikaylagawarecki/380/head 2025-12-04T08:53:58.7685151Z * [new branch] gh/mikaylagawarecki/380/orig -> origin/gh/mikaylagawarecki/380/orig 2025-12-04T08:53:58.7685243Z * [new branch] gh/mikaylagawarecki/381/base -> origin/gh/mikaylagawarecki/381/base 2025-12-04T08:53:58.7685370Z * [new branch] gh/mikaylagawarecki/381/head -> origin/gh/mikaylagawarecki/381/head 2025-12-04T08:53:58.7685463Z * [new branch] gh/mikaylagawarecki/381/orig -> origin/gh/mikaylagawarecki/381/orig 2025-12-04T08:53:58.7685556Z * [new branch] gh/mikaylagawarecki/382/base -> origin/gh/mikaylagawarecki/382/base 2025-12-04T08:53:58.7685648Z * [new branch] gh/mikaylagawarecki/382/head -> origin/gh/mikaylagawarecki/382/head 2025-12-04T08:53:58.7685740Z * [new branch] gh/mikaylagawarecki/382/orig -> origin/gh/mikaylagawarecki/382/orig 2025-12-04T08:53:58.7685839Z * [new branch] gh/mikaylagawarecki/383/base -> origin/gh/mikaylagawarecki/383/base 2025-12-04T08:53:58.7685934Z * [new branch] gh/mikaylagawarecki/383/head -> origin/gh/mikaylagawarecki/383/head 2025-12-04T08:53:58.7686027Z * [new branch] gh/mikaylagawarecki/383/orig -> origin/gh/mikaylagawarecki/383/orig 2025-12-04T08:53:58.7686122Z * [new branch] gh/mikaylagawarecki/384/base -> origin/gh/mikaylagawarecki/384/base 2025-12-04T08:53:58.7686214Z * [new branch] gh/mikaylagawarecki/384/head -> origin/gh/mikaylagawarecki/384/head 2025-12-04T08:53:58.7686306Z * [new branch] gh/mikaylagawarecki/384/orig -> origin/gh/mikaylagawarecki/384/orig 2025-12-04T08:53:58.7686402Z * [new branch] gh/mikaylagawarecki/385/base -> origin/gh/mikaylagawarecki/385/base 2025-12-04T08:53:58.7686496Z * [new branch] gh/mikaylagawarecki/385/head -> origin/gh/mikaylagawarecki/385/head 2025-12-04T08:53:58.7686593Z * [new branch] gh/mikaylagawarecki/385/orig -> origin/gh/mikaylagawarecki/385/orig 2025-12-04T08:53:58.7686688Z * [new branch] gh/mikaylagawarecki/386/base -> origin/gh/mikaylagawarecki/386/base 2025-12-04T08:53:58.7686814Z * [new branch] gh/mikaylagawarecki/386/head -> origin/gh/mikaylagawarecki/386/head 2025-12-04T08:53:58.7686913Z * [new branch] gh/mikaylagawarecki/386/orig -> origin/gh/mikaylagawarecki/386/orig 2025-12-04T08:53:58.7687006Z * [new branch] gh/mikaylagawarecki/387/base -> origin/gh/mikaylagawarecki/387/base 2025-12-04T08:53:58.7687099Z * [new branch] gh/mikaylagawarecki/387/head -> origin/gh/mikaylagawarecki/387/head 2025-12-04T08:53:58.7687191Z * [new branch] gh/mikaylagawarecki/387/orig -> origin/gh/mikaylagawarecki/387/orig 2025-12-04T08:53:58.7687283Z * [new branch] gh/mikaylagawarecki/388/base -> origin/gh/mikaylagawarecki/388/base 2025-12-04T08:53:58.7687413Z * [new branch] gh/mikaylagawarecki/388/head -> origin/gh/mikaylagawarecki/388/head 2025-12-04T08:53:58.7687507Z * [new branch] gh/mikaylagawarecki/388/orig -> origin/gh/mikaylagawarecki/388/orig 2025-12-04T08:53:58.7687599Z * [new branch] gh/mikaylagawarecki/389/base -> origin/gh/mikaylagawarecki/389/base 2025-12-04T08:53:58.7687693Z * [new branch] gh/mikaylagawarecki/389/head -> origin/gh/mikaylagawarecki/389/head 2025-12-04T08:53:58.7687786Z * [new branch] gh/mikaylagawarecki/389/orig -> origin/gh/mikaylagawarecki/389/orig 2025-12-04T08:53:58.7687879Z * [new branch] gh/mikaylagawarecki/390/base -> origin/gh/mikaylagawarecki/390/base 2025-12-04T08:53:58.7687972Z * [new branch] gh/mikaylagawarecki/390/head -> origin/gh/mikaylagawarecki/390/head 2025-12-04T08:53:58.7688065Z * [new branch] gh/mikaylagawarecki/390/orig -> origin/gh/mikaylagawarecki/390/orig 2025-12-04T08:53:58.7688160Z * [new branch] gh/mikaylagawarecki/391/base -> origin/gh/mikaylagawarecki/391/base 2025-12-04T08:53:58.7688254Z * [new branch] gh/mikaylagawarecki/391/head -> origin/gh/mikaylagawarecki/391/head 2025-12-04T08:53:58.7688349Z * [new branch] gh/mikaylagawarecki/391/orig -> origin/gh/mikaylagawarecki/391/orig 2025-12-04T08:53:58.7688480Z * [new branch] gh/mikaylagawarecki/392/base -> origin/gh/mikaylagawarecki/392/base 2025-12-04T08:53:58.7688575Z * [new branch] gh/mikaylagawarecki/392/head -> origin/gh/mikaylagawarecki/392/head 2025-12-04T08:53:58.7688669Z * [new branch] gh/mikaylagawarecki/392/orig -> origin/gh/mikaylagawarecki/392/orig 2025-12-04T08:53:58.7688743Z * [new branch] gh/mlazos/41/base -> origin/gh/mlazos/41/base 2025-12-04T08:53:58.7688815Z * [new branch] gh/mlazos/41/head -> origin/gh/mlazos/41/head 2025-12-04T08:53:58.7688886Z * [new branch] gh/mlazos/41/orig -> origin/gh/mlazos/41/orig 2025-12-04T08:53:58.7688955Z * [new branch] gh/mlazos/42/base -> origin/gh/mlazos/42/base 2025-12-04T08:53:58.7689022Z * [new branch] gh/mlazos/42/head -> origin/gh/mlazos/42/head 2025-12-04T08:53:58.7689090Z * [new branch] gh/mlazos/42/orig -> origin/gh/mlazos/42/orig 2025-12-04T08:53:58.7689156Z * [new branch] gh/mlazos/43/base -> origin/gh/mlazos/43/base 2025-12-04T08:53:58.7689223Z * [new branch] gh/mlazos/43/head -> origin/gh/mlazos/43/head 2025-12-04T08:53:58.7689291Z * [new branch] gh/mlazos/43/orig -> origin/gh/mlazos/43/orig 2025-12-04T08:53:58.7689358Z * [new branch] gh/mlazos/44/base -> origin/gh/mlazos/44/base 2025-12-04T08:53:58.7689428Z * [new branch] gh/mlazos/44/head -> origin/gh/mlazos/44/head 2025-12-04T08:53:58.7689497Z * [new branch] gh/mlazos/44/orig -> origin/gh/mlazos/44/orig 2025-12-04T08:53:58.7689566Z * [new branch] gh/mlazos/47/base -> origin/gh/mlazos/47/base 2025-12-04T08:53:58.7689634Z * [new branch] gh/mlazos/47/head -> origin/gh/mlazos/47/head 2025-12-04T08:53:58.7689701Z * [new branch] gh/mlazos/47/orig -> origin/gh/mlazos/47/orig 2025-12-04T08:53:58.7689770Z * [new branch] gh/mlazos/48/base -> origin/gh/mlazos/48/base 2025-12-04T08:53:58.7689839Z * [new branch] gh/mlazos/48/head -> origin/gh/mlazos/48/head 2025-12-04T08:53:58.7689906Z * [new branch] gh/mlazos/48/orig -> origin/gh/mlazos/48/orig 2025-12-04T08:53:58.7689974Z * [new branch] gh/mlazos/49/base -> origin/gh/mlazos/49/base 2025-12-04T08:53:58.7690045Z * [new branch] gh/mlazos/49/head -> origin/gh/mlazos/49/head 2025-12-04T08:53:58.7690141Z * [new branch] gh/mlazos/49/orig -> origin/gh/mlazos/49/orig 2025-12-04T08:53:58.7690210Z * [new branch] gh/mlazos/50/base -> origin/gh/mlazos/50/base 2025-12-04T08:53:58.7690278Z * [new branch] gh/mlazos/50/head -> origin/gh/mlazos/50/head 2025-12-04T08:53:58.7690345Z * [new branch] gh/mlazos/50/orig -> origin/gh/mlazos/50/orig 2025-12-04T08:53:58.7690414Z * [new branch] gh/mlazos/51/base -> origin/gh/mlazos/51/base 2025-12-04T08:53:58.7690482Z * [new branch] gh/mlazos/51/head -> origin/gh/mlazos/51/head 2025-12-04T08:53:58.7690553Z * [new branch] gh/mlazos/51/orig -> origin/gh/mlazos/51/orig 2025-12-04T08:53:58.7690623Z * [new branch] gh/mlazos/52/base -> origin/gh/mlazos/52/base 2025-12-04T08:53:58.7690689Z * [new branch] gh/mlazos/52/head -> origin/gh/mlazos/52/head 2025-12-04T08:53:58.7690758Z * [new branch] gh/mlazos/52/orig -> origin/gh/mlazos/52/orig 2025-12-04T08:53:58.7690826Z * [new branch] gh/mlazos/53/base -> origin/gh/mlazos/53/base 2025-12-04T08:53:58.7690892Z * [new branch] gh/mlazos/53/head -> origin/gh/mlazos/53/head 2025-12-04T08:53:58.7690958Z * [new branch] gh/mlazos/53/orig -> origin/gh/mlazos/53/orig 2025-12-04T08:53:58.7691050Z * [new branch] gh/mlazos/54/base -> origin/gh/mlazos/54/base 2025-12-04T08:53:58.7691117Z * [new branch] gh/mlazos/54/head -> origin/gh/mlazos/54/head 2025-12-04T08:53:58.7691183Z * [new branch] gh/mlazos/54/orig -> origin/gh/mlazos/54/orig 2025-12-04T08:53:58.7691251Z * [new branch] gh/mlazos/55/base -> origin/gh/mlazos/55/base 2025-12-04T08:53:58.7691318Z * [new branch] gh/mlazos/55/head -> origin/gh/mlazos/55/head 2025-12-04T08:53:58.7691386Z * [new branch] gh/mlazos/55/orig -> origin/gh/mlazos/55/orig 2025-12-04T08:53:58.7691459Z * [new branch] gh/mlazos/56/base -> origin/gh/mlazos/56/base 2025-12-04T08:53:58.7691525Z * [new branch] gh/mlazos/56/head -> origin/gh/mlazos/56/head 2025-12-04T08:53:58.7691591Z * [new branch] gh/mlazos/56/orig -> origin/gh/mlazos/56/orig 2025-12-04T08:53:58.7691661Z * [new branch] gh/mlazos/57/base -> origin/gh/mlazos/57/base 2025-12-04T08:53:58.7691729Z * [new branch] gh/mlazos/57/head -> origin/gh/mlazos/57/head 2025-12-04T08:53:58.7691797Z * [new branch] gh/mlazos/57/orig -> origin/gh/mlazos/57/orig 2025-12-04T08:53:58.7691865Z * [new branch] gh/mlazos/58/base -> origin/gh/mlazos/58/base 2025-12-04T08:53:58.7691931Z * [new branch] gh/mlazos/58/head -> origin/gh/mlazos/58/head 2025-12-04T08:53:58.7691999Z * [new branch] gh/mlazos/58/orig -> origin/gh/mlazos/58/orig 2025-12-04T08:53:58.7692069Z * [new branch] gh/mlazos/59/base -> origin/gh/mlazos/59/base 2025-12-04T08:53:58.7692136Z * [new branch] gh/mlazos/59/head -> origin/gh/mlazos/59/head 2025-12-04T08:53:58.7692207Z * [new branch] gh/mlazos/59/orig -> origin/gh/mlazos/59/orig 2025-12-04T08:53:58.7692274Z * [new branch] gh/mlazos/60/base -> origin/gh/mlazos/60/base 2025-12-04T08:53:58.7692341Z * [new branch] gh/mlazos/60/head -> origin/gh/mlazos/60/head 2025-12-04T08:53:58.7692410Z * [new branch] gh/mlazos/60/orig -> origin/gh/mlazos/60/orig 2025-12-04T08:53:58.7692476Z * [new branch] gh/mlazos/61/base -> origin/gh/mlazos/61/base 2025-12-04T08:53:58.7692544Z * [new branch] gh/mlazos/61/head -> origin/gh/mlazos/61/head 2025-12-04T08:53:58.7692614Z * [new branch] gh/mlazos/61/orig -> origin/gh/mlazos/61/orig 2025-12-04T08:53:58.7692708Z * [new branch] gh/mlazos/62/base -> origin/gh/mlazos/62/base 2025-12-04T08:53:58.7692775Z * [new branch] gh/mlazos/62/head -> origin/gh/mlazos/62/head 2025-12-04T08:53:58.7692843Z * [new branch] gh/mlazos/62/orig -> origin/gh/mlazos/62/orig 2025-12-04T08:53:58.7692913Z * [new branch] gh/mlazos/63/base -> origin/gh/mlazos/63/base 2025-12-04T08:53:58.7692979Z * [new branch] gh/mlazos/63/head -> origin/gh/mlazos/63/head 2025-12-04T08:53:58.7693048Z * [new branch] gh/mlazos/63/orig -> origin/gh/mlazos/63/orig 2025-12-04T08:53:58.7693117Z * [new branch] gh/mlazos/64/base -> origin/gh/mlazos/64/base 2025-12-04T08:53:58.7693183Z * [new branch] gh/mlazos/64/head -> origin/gh/mlazos/64/head 2025-12-04T08:53:58.7693253Z * [new branch] gh/mlazos/64/orig -> origin/gh/mlazos/64/orig 2025-12-04T08:53:58.7693323Z * [new branch] gh/mlazos/65/base -> origin/gh/mlazos/65/base 2025-12-04T08:53:58.7693393Z * [new branch] gh/mlazos/65/head -> origin/gh/mlazos/65/head 2025-12-04T08:53:58.7693459Z * [new branch] gh/mlazos/65/orig -> origin/gh/mlazos/65/orig 2025-12-04T08:53:58.7693550Z * [new branch] gh/mlazos/66/base -> origin/gh/mlazos/66/base 2025-12-04T08:53:58.7693621Z * [new branch] gh/mlazos/66/head -> origin/gh/mlazos/66/head 2025-12-04T08:53:58.7693689Z * [new branch] gh/mlazos/66/orig -> origin/gh/mlazos/66/orig 2025-12-04T08:53:58.7693757Z * [new branch] gh/mlazos/67/base -> origin/gh/mlazos/67/base 2025-12-04T08:53:58.7693825Z * [new branch] gh/mlazos/67/head -> origin/gh/mlazos/67/head 2025-12-04T08:53:58.7693891Z * [new branch] gh/mlazos/67/orig -> origin/gh/mlazos/67/orig 2025-12-04T08:53:58.7693958Z * [new branch] gh/mlazos/68/base -> origin/gh/mlazos/68/base 2025-12-04T08:53:58.7694025Z * [new branch] gh/mlazos/68/head -> origin/gh/mlazos/68/head 2025-12-04T08:53:58.7694091Z * [new branch] gh/mlazos/68/orig -> origin/gh/mlazos/68/orig 2025-12-04T08:53:58.7694159Z * [new branch] gh/mlazos/69/base -> origin/gh/mlazos/69/base 2025-12-04T08:53:58.7694230Z * [new branch] gh/mlazos/69/head -> origin/gh/mlazos/69/head 2025-12-04T08:53:58.7694296Z * [new branch] gh/mlazos/69/orig -> origin/gh/mlazos/69/orig 2025-12-04T08:53:58.7694363Z * [new branch] gh/mlazos/70/base -> origin/gh/mlazos/70/base 2025-12-04T08:53:58.7694432Z * [new branch] gh/mlazos/70/head -> origin/gh/mlazos/70/head 2025-12-04T08:53:58.7694499Z * [new branch] gh/mlazos/70/orig -> origin/gh/mlazos/70/orig 2025-12-04T08:53:58.7694569Z * [new branch] gh/mlazos/71/base -> origin/gh/mlazos/71/base 2025-12-04T08:53:58.7694638Z * [new branch] gh/mlazos/71/head -> origin/gh/mlazos/71/head 2025-12-04T08:53:58.7694705Z * [new branch] gh/mlazos/71/orig -> origin/gh/mlazos/71/orig 2025-12-04T08:53:58.7694771Z * [new branch] gh/mlazos/72/base -> origin/gh/mlazos/72/base 2025-12-04T08:53:58.7694842Z * [new branch] gh/mlazos/72/head -> origin/gh/mlazos/72/head 2025-12-04T08:53:58.7694910Z * [new branch] gh/mlazos/72/orig -> origin/gh/mlazos/72/orig 2025-12-04T08:53:58.7694980Z * [new branch] gh/mlazos/73/base -> origin/gh/mlazos/73/base 2025-12-04T08:53:58.7695047Z * [new branch] gh/mlazos/73/head -> origin/gh/mlazos/73/head 2025-12-04T08:53:58.7695114Z * [new branch] gh/mlazos/73/orig -> origin/gh/mlazos/73/orig 2025-12-04T08:53:58.7695222Z * [new branch] gh/mrmiywj/1/base -> origin/gh/mrmiywj/1/base 2025-12-04T08:53:58.7695289Z * [new branch] gh/mrmiywj/1/head -> origin/gh/mrmiywj/1/head 2025-12-04T08:53:58.7695366Z * [new branch] gh/muchulee8/73/base -> origin/gh/muchulee8/73/base 2025-12-04T08:53:58.7695440Z * [new branch] gh/muchulee8/73/head -> origin/gh/muchulee8/73/head 2025-12-04T08:53:58.7695515Z * [new branch] gh/muchulee8/73/orig -> origin/gh/muchulee8/73/orig 2025-12-04T08:53:58.7695604Z * [new branch] gh/naveenthangudu/1/base -> origin/gh/naveenthangudu/1/base 2025-12-04T08:53:58.7695691Z * [new branch] gh/naveenthangudu/1/head -> origin/gh/naveenthangudu/1/head 2025-12-04T08:53:58.7695771Z * [new branch] gh/naveenthangudu/1/orig -> origin/gh/naveenthangudu/1/orig 2025-12-04T08:53:58.7695852Z * [new branch] gh/naveenthangudu/2/base -> origin/gh/naveenthangudu/2/base 2025-12-04T08:53:58.7695936Z * [new branch] gh/naveenthangudu/2/head -> origin/gh/naveenthangudu/2/head 2025-12-04T08:53:58.7696017Z * [new branch] gh/naveenthangudu/2/orig -> origin/gh/naveenthangudu/2/orig 2025-12-04T08:53:58.7696099Z * [new branch] gh/naveenthangudu/3/base -> origin/gh/naveenthangudu/3/base 2025-12-04T08:53:58.7696208Z * [new branch] gh/naveenthangudu/3/head -> origin/gh/naveenthangudu/3/head 2025-12-04T08:53:58.7696291Z * [new branch] gh/naveenthangudu/3/orig -> origin/gh/naveenthangudu/3/orig 2025-12-04T08:53:58.7696372Z * [new branch] gh/naveenthangudu/4/base -> origin/gh/naveenthangudu/4/base 2025-12-04T08:53:58.7696453Z * [new branch] gh/naveenthangudu/4/head -> origin/gh/naveenthangudu/4/head 2025-12-04T08:53:58.7696532Z * [new branch] gh/naveenthangudu/4/orig -> origin/gh/naveenthangudu/4/orig 2025-12-04T08:53:58.7696613Z * [new branch] gh/naveenthangudu/5/base -> origin/gh/naveenthangudu/5/base 2025-12-04T08:53:58.7696693Z * [new branch] gh/naveenthangudu/5/head -> origin/gh/naveenthangudu/5/head 2025-12-04T08:53:58.7696893Z * [new branch] gh/naveenthangudu/5/orig -> origin/gh/naveenthangudu/5/orig 2025-12-04T08:53:58.7696975Z * [new branch] gh/naveenthangudu/6/base -> origin/gh/naveenthangudu/6/base 2025-12-04T08:53:58.7697057Z * [new branch] gh/naveenthangudu/6/head -> origin/gh/naveenthangudu/6/head 2025-12-04T08:53:58.7697138Z * [new branch] gh/naveenthangudu/6/orig -> origin/gh/naveenthangudu/6/orig 2025-12-04T08:53:58.7697219Z * [new branch] gh/naveenthangudu/7/base -> origin/gh/naveenthangudu/7/base 2025-12-04T08:53:58.7697298Z * [new branch] gh/naveenthangudu/7/head -> origin/gh/naveenthangudu/7/head 2025-12-04T08:53:58.7697378Z * [new branch] gh/naveenthangudu/7/orig -> origin/gh/naveenthangudu/7/orig 2025-12-04T08:53:58.7697462Z * [new branch] gh/naveenthangudu/8/base -> origin/gh/naveenthangudu/8/base 2025-12-04T08:53:58.7697540Z * [new branch] gh/naveenthangudu/8/head -> origin/gh/naveenthangudu/8/head 2025-12-04T08:53:58.7697619Z * [new branch] gh/naveenthangudu/8/orig -> origin/gh/naveenthangudu/8/orig 2025-12-04T08:53:58.7697702Z * [new branch] gh/naveenthangudu/9/base -> origin/gh/naveenthangudu/9/base 2025-12-04T08:53:58.7697782Z * [new branch] gh/naveenthangudu/9/head -> origin/gh/naveenthangudu/9/head 2025-12-04T08:53:58.7697864Z * [new branch] gh/naveenthangudu/9/orig -> origin/gh/naveenthangudu/9/orig 2025-12-04T08:53:58.7697941Z * [new branch] gh/nikitaved/1/base -> origin/gh/nikitaved/1/base 2025-12-04T08:53:58.7698016Z * [new branch] gh/nikitaved/1/head -> origin/gh/nikitaved/1/head 2025-12-04T08:53:58.7698088Z * [new branch] gh/nikitaved/1/orig -> origin/gh/nikitaved/1/orig 2025-12-04T08:53:58.7698203Z * [new branch] gh/nikitaved/10/base -> origin/gh/nikitaved/10/base 2025-12-04T08:53:58.7698276Z * [new branch] gh/nikitaved/10/head -> origin/gh/nikitaved/10/head 2025-12-04T08:53:58.7698349Z * [new branch] gh/nikitaved/10/orig -> origin/gh/nikitaved/10/orig 2025-12-04T08:53:58.7698422Z * [new branch] gh/nikitaved/11/base -> origin/gh/nikitaved/11/base 2025-12-04T08:53:58.7698494Z * [new branch] gh/nikitaved/11/head -> origin/gh/nikitaved/11/head 2025-12-04T08:53:58.7698569Z * [new branch] gh/nikitaved/11/orig -> origin/gh/nikitaved/11/orig 2025-12-04T08:53:58.7698639Z * [new branch] gh/nikitaved/12/base -> origin/gh/nikitaved/12/base 2025-12-04T08:53:58.7698712Z * [new branch] gh/nikitaved/12/head -> origin/gh/nikitaved/12/head 2025-12-04T08:53:58.7698785Z * [new branch] gh/nikitaved/12/orig -> origin/gh/nikitaved/12/orig 2025-12-04T08:53:58.7698860Z * [new branch] gh/nikitaved/13/base -> origin/gh/nikitaved/13/base 2025-12-04T08:53:58.7698932Z * [new branch] gh/nikitaved/13/head -> origin/gh/nikitaved/13/head 2025-12-04T08:53:58.7699006Z * [new branch] gh/nikitaved/13/orig -> origin/gh/nikitaved/13/orig 2025-12-04T08:53:58.7699119Z * [new branch] gh/nikitaved/14/base -> origin/gh/nikitaved/14/base 2025-12-04T08:53:58.7699190Z * [new branch] gh/nikitaved/14/head -> origin/gh/nikitaved/14/head 2025-12-04T08:53:58.7699264Z * [new branch] gh/nikitaved/14/orig -> origin/gh/nikitaved/14/orig 2025-12-04T08:53:58.7699336Z * [new branch] gh/nikitaved/15/base -> origin/gh/nikitaved/15/base 2025-12-04T08:53:58.7699408Z * [new branch] gh/nikitaved/15/head -> origin/gh/nikitaved/15/head 2025-12-04T08:53:58.7699480Z * [new branch] gh/nikitaved/15/orig -> origin/gh/nikitaved/15/orig 2025-12-04T08:53:58.7699553Z * [new branch] gh/nikitaved/16/base -> origin/gh/nikitaved/16/base 2025-12-04T08:53:58.7699624Z * [new branch] gh/nikitaved/16/head -> origin/gh/nikitaved/16/head 2025-12-04T08:53:58.7699697Z * [new branch] gh/nikitaved/16/orig -> origin/gh/nikitaved/16/orig 2025-12-04T08:53:58.7699771Z * [new branch] gh/nikitaved/2/base -> origin/gh/nikitaved/2/base 2025-12-04T08:53:58.7699844Z * [new branch] gh/nikitaved/2/head -> origin/gh/nikitaved/2/head 2025-12-04T08:53:58.7699915Z * [new branch] gh/nikitaved/2/orig -> origin/gh/nikitaved/2/orig 2025-12-04T08:53:58.7699986Z * [new branch] gh/nikitaved/4/base -> origin/gh/nikitaved/4/base 2025-12-04T08:53:58.7700059Z * [new branch] gh/nikitaved/4/head -> origin/gh/nikitaved/4/head 2025-12-04T08:53:58.7700128Z * [new branch] gh/nikitaved/4/orig -> origin/gh/nikitaved/4/orig 2025-12-04T08:53:58.7700202Z * [new branch] gh/nikitaved/5/base -> origin/gh/nikitaved/5/base 2025-12-04T08:53:58.7700272Z * [new branch] gh/nikitaved/5/head -> origin/gh/nikitaved/5/head 2025-12-04T08:53:58.7700342Z * [new branch] gh/nikitaved/5/orig -> origin/gh/nikitaved/5/orig 2025-12-04T08:53:58.7700414Z * [new branch] gh/nikitaved/6/base -> origin/gh/nikitaved/6/base 2025-12-04T08:53:58.7700486Z * [new branch] gh/nikitaved/6/head -> origin/gh/nikitaved/6/head 2025-12-04T08:53:58.7700558Z * [new branch] gh/nikitaved/6/orig -> origin/gh/nikitaved/6/orig 2025-12-04T08:53:58.7700628Z * [new branch] gh/nikitaved/8/base -> origin/gh/nikitaved/8/base 2025-12-04T08:53:58.7700700Z * [new branch] gh/nikitaved/8/head -> origin/gh/nikitaved/8/head 2025-12-04T08:53:58.7700770Z * [new branch] gh/nikitaved/8/orig -> origin/gh/nikitaved/8/orig 2025-12-04T08:53:58.7700871Z * [new branch] gh/nikitaved/9/base -> origin/gh/nikitaved/9/base 2025-12-04T08:53:58.7700945Z * [new branch] gh/nikitaved/9/head -> origin/gh/nikitaved/9/head 2025-12-04T08:53:58.7701017Z * [new branch] gh/nikitaved/9/orig -> origin/gh/nikitaved/9/orig 2025-12-04T08:53:58.7701086Z * [new branch] gh/oulgen/10/base -> origin/gh/oulgen/10/base 2025-12-04T08:53:58.7701157Z * [new branch] gh/oulgen/10/head -> origin/gh/oulgen/10/head 2025-12-04T08:53:58.7701226Z * [new branch] gh/oulgen/10/orig -> origin/gh/oulgen/10/orig 2025-12-04T08:53:58.7701295Z * [new branch] gh/oulgen/11/base -> origin/gh/oulgen/11/base 2025-12-04T08:53:58.7701363Z * [new branch] gh/oulgen/11/head -> origin/gh/oulgen/11/head 2025-12-04T08:53:58.7701431Z * [new branch] gh/oulgen/11/orig -> origin/gh/oulgen/11/orig 2025-12-04T08:53:58.7701499Z * [new branch] gh/oulgen/12/base -> origin/gh/oulgen/12/base 2025-12-04T08:53:58.7701568Z * [new branch] gh/oulgen/12/head -> origin/gh/oulgen/12/head 2025-12-04T08:53:58.7701635Z * [new branch] gh/oulgen/12/orig -> origin/gh/oulgen/12/orig 2025-12-04T08:53:58.7701735Z * [new branch] gh/oulgen/13/base -> origin/gh/oulgen/13/base 2025-12-04T08:53:58.7701802Z * [new branch] gh/oulgen/13/head -> origin/gh/oulgen/13/head 2025-12-04T08:53:58.7701870Z * [new branch] gh/oulgen/13/orig -> origin/gh/oulgen/13/orig 2025-12-04T08:53:58.7701939Z * [new branch] gh/oulgen/14/base -> origin/gh/oulgen/14/base 2025-12-04T08:53:58.7702007Z * [new branch] gh/oulgen/14/head -> origin/gh/oulgen/14/head 2025-12-04T08:53:58.7702074Z * [new branch] gh/oulgen/14/orig -> origin/gh/oulgen/14/orig 2025-12-04T08:53:58.7702145Z * [new branch] gh/oulgen/15/base -> origin/gh/oulgen/15/base 2025-12-04T08:53:58.7702213Z * [new branch] gh/oulgen/15/head -> origin/gh/oulgen/15/head 2025-12-04T08:53:58.7702281Z * [new branch] gh/oulgen/15/orig -> origin/gh/oulgen/15/orig 2025-12-04T08:53:58.7702354Z * [new branch] gh/oulgen/16/base -> origin/gh/oulgen/16/base 2025-12-04T08:53:58.7702421Z * [new branch] gh/oulgen/16/head -> origin/gh/oulgen/16/head 2025-12-04T08:53:58.7702487Z * [new branch] gh/oulgen/16/orig -> origin/gh/oulgen/16/orig 2025-12-04T08:53:58.7702558Z * [new branch] gh/oulgen/17/base -> origin/gh/oulgen/17/base 2025-12-04T08:53:58.7702624Z * [new branch] gh/oulgen/17/head -> origin/gh/oulgen/17/head 2025-12-04T08:53:58.7702692Z * [new branch] gh/oulgen/17/orig -> origin/gh/oulgen/17/orig 2025-12-04T08:53:58.7702765Z * [new branch] gh/oulgen/18/base -> origin/gh/oulgen/18/base 2025-12-04T08:53:58.7702831Z * [new branch] gh/oulgen/18/head -> origin/gh/oulgen/18/head 2025-12-04T08:53:58.7702900Z * [new branch] gh/oulgen/18/orig -> origin/gh/oulgen/18/orig 2025-12-04T08:53:58.7702968Z * [new branch] gh/oulgen/19/base -> origin/gh/oulgen/19/base 2025-12-04T08:53:58.7703036Z * [new branch] gh/oulgen/19/head -> origin/gh/oulgen/19/head 2025-12-04T08:53:58.7703104Z * [new branch] gh/oulgen/19/orig -> origin/gh/oulgen/19/orig 2025-12-04T08:53:58.7703170Z * [new branch] gh/oulgen/20/base -> origin/gh/oulgen/20/base 2025-12-04T08:53:58.7703236Z * [new branch] gh/oulgen/20/head -> origin/gh/oulgen/20/head 2025-12-04T08:53:58.7703305Z * [new branch] gh/oulgen/20/orig -> origin/gh/oulgen/20/orig 2025-12-04T08:53:58.7703403Z * [new branch] gh/oulgen/21/base -> origin/gh/oulgen/21/base 2025-12-04T08:53:58.7703471Z * [new branch] gh/oulgen/21/head -> origin/gh/oulgen/21/head 2025-12-04T08:53:58.7703540Z * [new branch] gh/oulgen/21/orig -> origin/gh/oulgen/21/orig 2025-12-04T08:53:58.7703605Z * [new branch] gh/oulgen/22/base -> origin/gh/oulgen/22/base 2025-12-04T08:53:58.7703675Z * [new branch] gh/oulgen/22/head -> origin/gh/oulgen/22/head 2025-12-04T08:53:58.7703742Z * [new branch] gh/oulgen/22/orig -> origin/gh/oulgen/22/orig 2025-12-04T08:53:58.7703809Z * [new branch] gh/oulgen/23/base -> origin/gh/oulgen/23/base 2025-12-04T08:53:58.7703877Z * [new branch] gh/oulgen/23/head -> origin/gh/oulgen/23/head 2025-12-04T08:53:58.7703944Z * [new branch] gh/oulgen/23/orig -> origin/gh/oulgen/23/orig 2025-12-04T08:53:58.7704012Z * [new branch] gh/oulgen/24/base -> origin/gh/oulgen/24/base 2025-12-04T08:53:58.7704080Z * [new branch] gh/oulgen/24/head -> origin/gh/oulgen/24/head 2025-12-04T08:53:58.7704149Z * [new branch] gh/oulgen/24/orig -> origin/gh/oulgen/24/orig 2025-12-04T08:53:58.7704216Z * [new branch] gh/oulgen/25/base -> origin/gh/oulgen/25/base 2025-12-04T08:53:58.7704308Z * [new branch] gh/oulgen/25/head -> origin/gh/oulgen/25/head 2025-12-04T08:53:58.7704376Z * [new branch] gh/oulgen/25/orig -> origin/gh/oulgen/25/orig 2025-12-04T08:53:58.7704442Z * [new branch] gh/oulgen/26/base -> origin/gh/oulgen/26/base 2025-12-04T08:53:58.7704509Z * [new branch] gh/oulgen/26/head -> origin/gh/oulgen/26/head 2025-12-04T08:53:58.7704576Z * [new branch] gh/oulgen/26/orig -> origin/gh/oulgen/26/orig 2025-12-04T08:53:58.7704646Z * [new branch] gh/oulgen/4/base -> origin/gh/oulgen/4/base 2025-12-04T08:53:58.7704716Z * [new branch] gh/oulgen/4/head -> origin/gh/oulgen/4/head 2025-12-04T08:53:58.7704782Z * [new branch] gh/oulgen/4/orig -> origin/gh/oulgen/4/orig 2025-12-04T08:53:58.7704848Z * [new branch] gh/oulgen/7/base -> origin/gh/oulgen/7/base 2025-12-04T08:53:58.7704917Z * [new branch] gh/oulgen/7/head -> origin/gh/oulgen/7/head 2025-12-04T08:53:58.7704984Z * [new branch] gh/oulgen/7/orig -> origin/gh/oulgen/7/orig 2025-12-04T08:53:58.7705051Z * [new branch] gh/oulgen/8/base -> origin/gh/oulgen/8/base 2025-12-04T08:53:58.7705117Z * [new branch] gh/oulgen/8/head -> origin/gh/oulgen/8/head 2025-12-04T08:53:58.7705183Z * [new branch] gh/oulgen/8/orig -> origin/gh/oulgen/8/orig 2025-12-04T08:53:58.7705248Z * [new branch] gh/oulgen/9/base -> origin/gh/oulgen/9/base 2025-12-04T08:53:58.7705316Z * [new branch] gh/oulgen/9/head -> origin/gh/oulgen/9/head 2025-12-04T08:53:58.7705382Z * [new branch] gh/oulgen/9/orig -> origin/gh/oulgen/9/orig 2025-12-04T08:53:58.7705488Z * [new branch] gh/patvig/mtia-serialization -> origin/gh/patvig/mtia-serialization 2025-12-04T08:53:58.7705559Z * [new branch] gh/pearu/108/base -> origin/gh/pearu/108/base 2025-12-04T08:53:58.7705627Z * [new branch] gh/pearu/108/head -> origin/gh/pearu/108/head 2025-12-04T08:53:58.7705696Z * [new branch] gh/pearu/108/orig -> origin/gh/pearu/108/orig 2025-12-04T08:53:58.7705764Z * [new branch] gh/pearu/109/base -> origin/gh/pearu/109/base 2025-12-04T08:53:58.7705831Z * [new branch] gh/pearu/109/head -> origin/gh/pearu/109/head 2025-12-04T08:53:58.7705899Z * [new branch] gh/pearu/109/orig -> origin/gh/pearu/109/orig 2025-12-04T08:53:58.7705995Z * [new branch] gh/pearu/110/base -> origin/gh/pearu/110/base 2025-12-04T08:53:58.7706063Z * [new branch] gh/pearu/110/head -> origin/gh/pearu/110/head 2025-12-04T08:53:58.7706132Z * [new branch] gh/pearu/110/orig -> origin/gh/pearu/110/orig 2025-12-04T08:53:58.7706201Z * [new branch] gh/pearu/111/base -> origin/gh/pearu/111/base 2025-12-04T08:53:58.7706269Z * [new branch] gh/pearu/111/head -> origin/gh/pearu/111/head 2025-12-04T08:53:58.7706338Z * [new branch] gh/pearu/111/orig -> origin/gh/pearu/111/orig 2025-12-04T08:53:58.7706405Z * [new branch] gh/pearu/112/base -> origin/gh/pearu/112/base 2025-12-04T08:53:58.7706473Z * [new branch] gh/pearu/112/head -> origin/gh/pearu/112/head 2025-12-04T08:53:58.7706542Z * [new branch] gh/pearu/112/orig -> origin/gh/pearu/112/orig 2025-12-04T08:53:58.7706610Z * [new branch] gh/pearu/115/base -> origin/gh/pearu/115/base 2025-12-04T08:53:58.7706678Z * [new branch] gh/pearu/115/head -> origin/gh/pearu/115/head 2025-12-04T08:53:58.7706793Z * [new branch] gh/pearu/115/orig -> origin/gh/pearu/115/orig 2025-12-04T08:53:58.7706900Z * [new branch] gh/pearu/116/base -> origin/gh/pearu/116/base 2025-12-04T08:53:58.7706967Z * [new branch] gh/pearu/116/head -> origin/gh/pearu/116/head 2025-12-04T08:53:58.7707035Z * [new branch] gh/pearu/116/orig -> origin/gh/pearu/116/orig 2025-12-04T08:53:58.7707102Z * [new branch] gh/pearu/117/base -> origin/gh/pearu/117/base 2025-12-04T08:53:58.7707171Z * [new branch] gh/pearu/117/head -> origin/gh/pearu/117/head 2025-12-04T08:53:58.7707239Z * [new branch] gh/pearu/117/orig -> origin/gh/pearu/117/orig 2025-12-04T08:53:58.7707308Z * [new branch] gh/pearu/118/base -> origin/gh/pearu/118/base 2025-12-04T08:53:58.7707376Z * [new branch] gh/pearu/118/head -> origin/gh/pearu/118/head 2025-12-04T08:53:58.7707443Z * [new branch] gh/pearu/118/orig -> origin/gh/pearu/118/orig 2025-12-04T08:53:58.7707511Z * [new branch] gh/pearu/119/base -> origin/gh/pearu/119/base 2025-12-04T08:53:58.7707579Z * [new branch] gh/pearu/119/head -> origin/gh/pearu/119/head 2025-12-04T08:53:58.7707647Z * [new branch] gh/pearu/119/orig -> origin/gh/pearu/119/orig 2025-12-04T08:53:58.7707716Z * [new branch] gh/pearu/139/base -> origin/gh/pearu/139/base 2025-12-04T08:53:58.7707784Z * [new branch] gh/pearu/139/head -> origin/gh/pearu/139/head 2025-12-04T08:53:58.7707851Z * [new branch] gh/pearu/139/orig -> origin/gh/pearu/139/orig 2025-12-04T08:53:58.7707919Z * [new branch] gh/pearu/140/base -> origin/gh/pearu/140/base 2025-12-04T08:53:58.7707988Z * [new branch] gh/pearu/140/head -> origin/gh/pearu/140/head 2025-12-04T08:53:58.7708055Z * [new branch] gh/pearu/140/orig -> origin/gh/pearu/140/orig 2025-12-04T08:53:58.7708123Z * [new branch] gh/pearu/142/base -> origin/gh/pearu/142/base 2025-12-04T08:53:58.7708194Z * [new branch] gh/pearu/142/head -> origin/gh/pearu/142/head 2025-12-04T08:53:58.7708262Z * [new branch] gh/pearu/142/orig -> origin/gh/pearu/142/orig 2025-12-04T08:53:58.7708329Z * [new branch] gh/pearu/143/base -> origin/gh/pearu/143/base 2025-12-04T08:53:58.7708397Z * [new branch] gh/pearu/143/head -> origin/gh/pearu/143/head 2025-12-04T08:53:58.7708464Z * [new branch] gh/pearu/143/orig -> origin/gh/pearu/143/orig 2025-12-04T08:53:58.7708570Z * [new branch] gh/pearu/147/base -> origin/gh/pearu/147/base 2025-12-04T08:53:58.7708640Z * [new branch] gh/pearu/147/head -> origin/gh/pearu/147/head 2025-12-04T08:53:58.7708708Z * [new branch] gh/pearu/147/orig -> origin/gh/pearu/147/orig 2025-12-04T08:53:58.7708776Z * [new branch] gh/pearu/149/base -> origin/gh/pearu/149/base 2025-12-04T08:53:58.7708845Z * [new branch] gh/pearu/149/head -> origin/gh/pearu/149/head 2025-12-04T08:53:58.7708913Z * [new branch] gh/pearu/149/orig -> origin/gh/pearu/149/orig 2025-12-04T08:53:58.7708982Z * [new branch] gh/pearu/150/base -> origin/gh/pearu/150/base 2025-12-04T08:53:58.7709050Z * [new branch] gh/pearu/150/head -> origin/gh/pearu/150/head 2025-12-04T08:53:58.7709118Z * [new branch] gh/pearu/150/orig -> origin/gh/pearu/150/orig 2025-12-04T08:53:58.7709188Z * [new branch] gh/pearu/151/base -> origin/gh/pearu/151/base 2025-12-04T08:53:58.7709256Z * [new branch] gh/pearu/151/head -> origin/gh/pearu/151/head 2025-12-04T08:53:58.7709322Z * [new branch] gh/pearu/151/orig -> origin/gh/pearu/151/orig 2025-12-04T08:53:58.7709391Z * [new branch] gh/pearu/152/base -> origin/gh/pearu/152/base 2025-12-04T08:53:58.7709484Z * [new branch] gh/pearu/152/head -> origin/gh/pearu/152/head 2025-12-04T08:53:58.7709551Z * [new branch] gh/pearu/152/orig -> origin/gh/pearu/152/orig 2025-12-04T08:53:58.7709620Z * [new branch] gh/pearu/153/base -> origin/gh/pearu/153/base 2025-12-04T08:53:58.7709688Z * [new branch] gh/pearu/153/head -> origin/gh/pearu/153/head 2025-12-04T08:53:58.7709756Z * [new branch] gh/pearu/153/orig -> origin/gh/pearu/153/orig 2025-12-04T08:53:58.7709824Z * [new branch] gh/pearu/154/base -> origin/gh/pearu/154/base 2025-12-04T08:53:58.7709893Z * [new branch] gh/pearu/154/head -> origin/gh/pearu/154/head 2025-12-04T08:53:58.7709960Z * [new branch] gh/pearu/154/orig -> origin/gh/pearu/154/orig 2025-12-04T08:53:58.7710028Z * [new branch] gh/pearu/155/base -> origin/gh/pearu/155/base 2025-12-04T08:53:58.7710097Z * [new branch] gh/pearu/155/head -> origin/gh/pearu/155/head 2025-12-04T08:53:58.7710166Z * [new branch] gh/pearu/155/orig -> origin/gh/pearu/155/orig 2025-12-04T08:53:58.7710235Z * [new branch] gh/pearu/156/base -> origin/gh/pearu/156/base 2025-12-04T08:53:58.7710301Z * [new branch] gh/pearu/156/head -> origin/gh/pearu/156/head 2025-12-04T08:53:58.7710369Z * [new branch] gh/pearu/156/orig -> origin/gh/pearu/156/orig 2025-12-04T08:53:58.7710436Z * [new branch] gh/pearu/56/base -> origin/gh/pearu/56/base 2025-12-04T08:53:58.7710505Z * [new branch] gh/pearu/56/head -> origin/gh/pearu/56/head 2025-12-04T08:53:58.7710572Z * [new branch] gh/pearu/56/orig -> origin/gh/pearu/56/orig 2025-12-04T08:53:58.7710638Z * [new branch] gh/pearu/97/base -> origin/gh/pearu/97/base 2025-12-04T08:53:58.7710705Z * [new branch] gh/pearu/97/head -> origin/gh/pearu/97/head 2025-12-04T08:53:58.7710772Z * [new branch] gh/pearu/97/orig -> origin/gh/pearu/97/orig 2025-12-04T08:53:58.7710850Z * [new branch] gh/pianpwk/21/base -> origin/gh/pianpwk/21/base 2025-12-04T08:53:58.7710921Z * [new branch] gh/pianpwk/21/head -> origin/gh/pianpwk/21/head 2025-12-04T08:53:58.7710993Z * [new branch] gh/pianpwk/28/base -> origin/gh/pianpwk/28/base 2025-12-04T08:53:58.7711063Z * [new branch] gh/pianpwk/28/head -> origin/gh/pianpwk/28/head 2025-12-04T08:53:58.7711162Z * [new branch] gh/pianpwk/28/orig -> origin/gh/pianpwk/28/orig 2025-12-04T08:53:58.7711234Z * [new branch] gh/pianpwk/29/base -> origin/gh/pianpwk/29/base 2025-12-04T08:53:58.7711304Z * [new branch] gh/pianpwk/29/head -> origin/gh/pianpwk/29/head 2025-12-04T08:53:58.7711374Z * [new branch] gh/pianpwk/29/orig -> origin/gh/pianpwk/29/orig 2025-12-04T08:53:58.7711446Z * [new branch] gh/pianpwk/30/base -> origin/gh/pianpwk/30/base 2025-12-04T08:53:58.7711516Z * [new branch] gh/pianpwk/30/head -> origin/gh/pianpwk/30/head 2025-12-04T08:53:58.7711585Z * [new branch] gh/pianpwk/30/orig -> origin/gh/pianpwk/30/orig 2025-12-04T08:53:58.7711657Z * [new branch] gh/pianpwk/31/base -> origin/gh/pianpwk/31/base 2025-12-04T08:53:58.7711728Z * [new branch] gh/pianpwk/31/head -> origin/gh/pianpwk/31/head 2025-12-04T08:53:58.7711799Z * [new branch] gh/pianpwk/31/orig -> origin/gh/pianpwk/31/orig 2025-12-04T08:53:58.7711871Z * [new branch] gh/pianpwk/32/base -> origin/gh/pianpwk/32/base 2025-12-04T08:53:58.7711941Z * [new branch] gh/pianpwk/32/head -> origin/gh/pianpwk/32/head 2025-12-04T08:53:58.7712036Z * [new branch] gh/pianpwk/32/orig -> origin/gh/pianpwk/32/orig 2025-12-04T08:53:58.7712107Z * [new branch] gh/pianpwk/33/base -> origin/gh/pianpwk/33/base 2025-12-04T08:53:58.7712176Z * [new branch] gh/pianpwk/33/head -> origin/gh/pianpwk/33/head 2025-12-04T08:53:58.7712248Z * [new branch] gh/pianpwk/33/orig -> origin/gh/pianpwk/33/orig 2025-12-04T08:53:58.7712316Z * [new branch] gh/pianpwk/34/base -> origin/gh/pianpwk/34/base 2025-12-04T08:53:58.7712385Z * [new branch] gh/pianpwk/34/head -> origin/gh/pianpwk/34/head 2025-12-04T08:53:58.7712459Z * [new branch] gh/pianpwk/34/orig -> origin/gh/pianpwk/34/orig 2025-12-04T08:53:58.7712528Z * [new branch] gh/pianpwk/35/base -> origin/gh/pianpwk/35/base 2025-12-04T08:53:58.7712597Z * [new branch] gh/pianpwk/35/head -> origin/gh/pianpwk/35/head 2025-12-04T08:53:58.7712669Z * [new branch] gh/pianpwk/35/orig -> origin/gh/pianpwk/35/orig 2025-12-04T08:53:58.7712736Z * [new branch] gh/rec/141/base -> origin/gh/rec/141/base 2025-12-04T08:53:58.7712801Z * [new branch] gh/rec/141/head -> origin/gh/rec/141/head 2025-12-04T08:53:58.7712869Z * [new branch] gh/rec/153/base -> origin/gh/rec/153/base 2025-12-04T08:53:58.7712933Z * [new branch] gh/rec/153/head -> origin/gh/rec/153/head 2025-12-04T08:53:58.7712996Z * [new branch] gh/rec/153/orig -> origin/gh/rec/153/orig 2025-12-04T08:53:58.7713063Z * [new branch] gh/rec/154/base -> origin/gh/rec/154/base 2025-12-04T08:53:58.7713126Z * [new branch] gh/rec/154/head -> origin/gh/rec/154/head 2025-12-04T08:53:58.7713190Z * [new branch] gh/rec/154/orig -> origin/gh/rec/154/orig 2025-12-04T08:53:58.7713256Z * [new branch] gh/rec/164/base -> origin/gh/rec/164/base 2025-12-04T08:53:58.7713321Z * [new branch] gh/rec/164/head -> origin/gh/rec/164/head 2025-12-04T08:53:58.7713384Z * [new branch] gh/rec/164/orig -> origin/gh/rec/164/orig 2025-12-04T08:53:58.7713449Z * [new branch] gh/rec/166/base -> origin/gh/rec/166/base 2025-12-04T08:53:58.7713513Z * [new branch] gh/rec/166/head -> origin/gh/rec/166/head 2025-12-04T08:53:58.7713576Z * [new branch] gh/rec/166/orig -> origin/gh/rec/166/orig 2025-12-04T08:53:58.7713665Z * [new branch] gh/rec/167/base -> origin/gh/rec/167/base 2025-12-04T08:53:58.7713730Z * [new branch] gh/rec/167/head -> origin/gh/rec/167/head 2025-12-04T08:53:58.7713794Z * [new branch] gh/rec/167/orig -> origin/gh/rec/167/orig 2025-12-04T08:53:58.7713859Z * [new branch] gh/rec/168/base -> origin/gh/rec/168/base 2025-12-04T08:53:58.7713924Z * [new branch] gh/rec/168/head -> origin/gh/rec/168/head 2025-12-04T08:53:58.7713988Z * [new branch] gh/rec/168/orig -> origin/gh/rec/168/orig 2025-12-04T08:53:58.7714053Z * [new branch] gh/rec/169/base -> origin/gh/rec/169/base 2025-12-04T08:53:58.7714116Z * [new branch] gh/rec/169/head -> origin/gh/rec/169/head 2025-12-04T08:53:58.7714181Z * [new branch] gh/rec/169/orig -> origin/gh/rec/169/orig 2025-12-04T08:53:58.7714245Z * [new branch] gh/rec/170/base -> origin/gh/rec/170/base 2025-12-04T08:53:58.7714309Z * [new branch] gh/rec/170/head -> origin/gh/rec/170/head 2025-12-04T08:53:58.7714372Z * [new branch] gh/rec/170/orig -> origin/gh/rec/170/orig 2025-12-04T08:53:58.7714435Z * [new branch] gh/rec/171/base -> origin/gh/rec/171/base 2025-12-04T08:53:58.7714533Z * [new branch] gh/rec/171/head -> origin/gh/rec/171/head 2025-12-04T08:53:58.7714600Z * [new branch] gh/rec/171/orig -> origin/gh/rec/171/orig 2025-12-04T08:53:58.7714664Z * [new branch] gh/rec/172/base -> origin/gh/rec/172/base 2025-12-04T08:53:58.7714728Z * [new branch] gh/rec/172/head -> origin/gh/rec/172/head 2025-12-04T08:53:58.7714793Z * [new branch] gh/rec/172/orig -> origin/gh/rec/172/orig 2025-12-04T08:53:58.7714857Z * [new branch] gh/rec/173/base -> origin/gh/rec/173/base 2025-12-04T08:53:58.7714922Z * [new branch] gh/rec/173/head -> origin/gh/rec/173/head 2025-12-04T08:53:58.7714986Z * [new branch] gh/rec/173/orig -> origin/gh/rec/173/orig 2025-12-04T08:53:58.7715048Z * [new branch] gh/rec/174/base -> origin/gh/rec/174/base 2025-12-04T08:53:58.7715112Z * [new branch] gh/rec/174/head -> origin/gh/rec/174/head 2025-12-04T08:53:58.7715179Z * [new branch] gh/rec/174/orig -> origin/gh/rec/174/orig 2025-12-04T08:53:58.7715242Z * [new branch] gh/rec/175/base -> origin/gh/rec/175/base 2025-12-04T08:53:58.7715306Z * [new branch] gh/rec/175/head -> origin/gh/rec/175/head 2025-12-04T08:53:58.7715371Z * [new branch] gh/rec/175/orig -> origin/gh/rec/175/orig 2025-12-04T08:53:58.7715435Z * [new branch] gh/rec/176/base -> origin/gh/rec/176/base 2025-12-04T08:53:58.7715500Z * [new branch] gh/rec/176/head -> origin/gh/rec/176/head 2025-12-04T08:53:58.7715566Z * [new branch] gh/rec/176/orig -> origin/gh/rec/176/orig 2025-12-04T08:53:58.7715628Z * [new branch] gh/rec/177/base -> origin/gh/rec/177/base 2025-12-04T08:53:58.7715694Z * [new branch] gh/rec/177/head -> origin/gh/rec/177/head 2025-12-04T08:53:58.7715759Z * [new branch] gh/rec/177/orig -> origin/gh/rec/177/orig 2025-12-04T08:53:58.7715848Z * [new branch] gh/robert-hardwick/3/base -> origin/gh/robert-hardwick/3/base 2025-12-04T08:53:58.7715939Z * [new branch] gh/robert-hardwick/3/head -> origin/gh/robert-hardwick/3/head 2025-12-04T08:53:58.7716022Z * [new branch] gh/robert-hardwick/3/orig -> origin/gh/robert-hardwick/3/orig 2025-12-04T08:53:58.7716106Z * [new branch] gh/robert-hardwick/4/base -> origin/gh/robert-hardwick/4/base 2025-12-04T08:53:58.7716215Z * [new branch] gh/robert-hardwick/4/head -> origin/gh/robert-hardwick/4/head 2025-12-04T08:53:58.7716296Z * [new branch] gh/robert-hardwick/4/orig -> origin/gh/robert-hardwick/4/orig 2025-12-04T08:53:58.7716379Z * [new branch] gh/robert-hardwick/5/base -> origin/gh/robert-hardwick/5/base 2025-12-04T08:53:58.7716463Z * [new branch] gh/robert-hardwick/5/head -> origin/gh/robert-hardwick/5/head 2025-12-04T08:53:58.7716546Z * [new branch] gh/robert-hardwick/5/orig -> origin/gh/robert-hardwick/5/orig 2025-12-04T08:53:58.7716627Z * [new branch] gh/robert-hardwick/6/base -> origin/gh/robert-hardwick/6/base 2025-12-04T08:53:58.7716712Z * [new branch] gh/robert-hardwick/6/head -> origin/gh/robert-hardwick/6/head 2025-12-04T08:53:58.7716843Z * [new branch] gh/robert-hardwick/6/orig -> origin/gh/robert-hardwick/6/orig 2025-12-04T08:53:58.7716927Z * [new branch] gh/robert-hardwick/7/base -> origin/gh/robert-hardwick/7/base 2025-12-04T08:53:58.7717012Z * [new branch] gh/robert-hardwick/7/head -> origin/gh/robert-hardwick/7/head 2025-12-04T08:53:58.7717094Z * [new branch] gh/robert-hardwick/7/orig -> origin/gh/robert-hardwick/7/orig 2025-12-04T08:53:58.7717176Z * [new branch] gh/robert-hardwick/8/base -> origin/gh/robert-hardwick/8/base 2025-12-04T08:53:58.7717300Z * [new branch] gh/robert-hardwick/8/head -> origin/gh/robert-hardwick/8/head 2025-12-04T08:53:58.7717384Z * [new branch] gh/robert-hardwick/8/orig -> origin/gh/robert-hardwick/8/orig 2025-12-04T08:53:58.7717467Z * [new branch] gh/robert-hardwick/9/base -> origin/gh/robert-hardwick/9/base 2025-12-04T08:53:58.7717549Z * [new branch] gh/robert-hardwick/9/head -> origin/gh/robert-hardwick/9/head 2025-12-04T08:53:58.7717632Z * [new branch] gh/robert-hardwick/9/orig -> origin/gh/robert-hardwick/9/orig 2025-12-04T08:53:58.7717707Z * [new branch] gh/rtimpe/1/base -> origin/gh/rtimpe/1/base 2025-12-04T08:53:58.7717776Z * [new branch] gh/rtimpe/1/head -> origin/gh/rtimpe/1/head 2025-12-04T08:53:58.7717845Z * [new branch] gh/rtimpe/2/base -> origin/gh/rtimpe/2/base 2025-12-04T08:53:58.7717915Z * [new branch] gh/rtimpe/2/head -> origin/gh/rtimpe/2/head 2025-12-04T08:53:58.7717987Z * [new branch] gh/rtimpe/22/base -> origin/gh/rtimpe/22/base 2025-12-04T08:53:58.7718056Z * [new branch] gh/rtimpe/22/head -> origin/gh/rtimpe/22/head 2025-12-04T08:53:58.7718125Z * [new branch] gh/rtimpe/22/orig -> origin/gh/rtimpe/22/orig 2025-12-04T08:53:58.7718193Z * [new branch] gh/rtimpe/23/base -> origin/gh/rtimpe/23/base 2025-12-04T08:53:58.7718260Z * [new branch] gh/rtimpe/23/head -> origin/gh/rtimpe/23/head 2025-12-04T08:53:58.7718330Z * [new branch] gh/rtimpe/23/orig -> origin/gh/rtimpe/23/orig 2025-12-04T08:53:58.7718397Z * [new branch] gh/rtimpe/24/base -> origin/gh/rtimpe/24/base 2025-12-04T08:53:58.7718463Z * [new branch] gh/rtimpe/24/head -> origin/gh/rtimpe/24/head 2025-12-04T08:53:58.7718531Z * [new branch] gh/rtimpe/24/orig -> origin/gh/rtimpe/24/orig 2025-12-04T08:53:58.7718600Z * [new branch] gh/rtimpe/25/base -> origin/gh/rtimpe/25/base 2025-12-04T08:53:58.7718666Z * [new branch] gh/rtimpe/25/head -> origin/gh/rtimpe/25/head 2025-12-04T08:53:58.7718734Z * [new branch] gh/rtimpe/25/orig -> origin/gh/rtimpe/25/orig 2025-12-04T08:53:58.7718803Z * [new branch] gh/rtimpe/26/base -> origin/gh/rtimpe/26/base 2025-12-04T08:53:58.7718871Z * [new branch] gh/rtimpe/26/head -> origin/gh/rtimpe/26/head 2025-12-04T08:53:58.7718986Z * [new branch] gh/rtimpe/26/orig -> origin/gh/rtimpe/26/orig 2025-12-04T08:53:58.7719053Z * [new branch] gh/rtimpe/27/base -> origin/gh/rtimpe/27/base 2025-12-04T08:53:58.7719122Z * [new branch] gh/rtimpe/27/head -> origin/gh/rtimpe/27/head 2025-12-04T08:53:58.7719191Z * [new branch] gh/rtimpe/27/orig -> origin/gh/rtimpe/27/orig 2025-12-04T08:53:58.7719260Z * [new branch] gh/rtimpe/28/base -> origin/gh/rtimpe/28/base 2025-12-04T08:53:58.7719328Z * [new branch] gh/rtimpe/28/head -> origin/gh/rtimpe/28/head 2025-12-04T08:53:58.7719397Z * [new branch] gh/rtimpe/28/orig -> origin/gh/rtimpe/28/orig 2025-12-04T08:53:58.7719464Z * [new branch] gh/rtimpe/29/base -> origin/gh/rtimpe/29/base 2025-12-04T08:53:58.7719532Z * [new branch] gh/rtimpe/29/head -> origin/gh/rtimpe/29/head 2025-12-04T08:53:58.7719601Z * [new branch] gh/rtimpe/29/orig -> origin/gh/rtimpe/29/orig 2025-12-04T08:53:58.7719670Z * [new branch] gh/rtimpe/3/base -> origin/gh/rtimpe/3/base 2025-12-04T08:53:58.7719740Z * [new branch] gh/rtimpe/3/head -> origin/gh/rtimpe/3/head 2025-12-04T08:53:58.7719807Z * [new branch] gh/rtimpe/30/base -> origin/gh/rtimpe/30/base 2025-12-04T08:53:58.7719898Z * [new branch] gh/rtimpe/30/head -> origin/gh/rtimpe/30/head 2025-12-04T08:53:58.7719969Z * [new branch] gh/rtimpe/30/orig -> origin/gh/rtimpe/30/orig 2025-12-04T08:53:58.7720035Z * [new branch] gh/rtimpe/31/base -> origin/gh/rtimpe/31/base 2025-12-04T08:53:58.7720102Z * [new branch] gh/rtimpe/31/head -> origin/gh/rtimpe/31/head 2025-12-04T08:53:58.7720172Z * [new branch] gh/rtimpe/31/orig -> origin/gh/rtimpe/31/orig 2025-12-04T08:53:58.7720238Z * [new branch] gh/rtimpe/32/base -> origin/gh/rtimpe/32/base 2025-12-04T08:53:58.7720308Z * [new branch] gh/rtimpe/32/head -> origin/gh/rtimpe/32/head 2025-12-04T08:53:58.7720377Z * [new branch] gh/rtimpe/32/orig -> origin/gh/rtimpe/32/orig 2025-12-04T08:53:58.7720445Z * [new branch] gh/rtimpe/33/base -> origin/gh/rtimpe/33/base 2025-12-04T08:53:58.7720514Z * [new branch] gh/rtimpe/33/head -> origin/gh/rtimpe/33/head 2025-12-04T08:53:58.7720582Z * [new branch] gh/rtimpe/33/orig -> origin/gh/rtimpe/33/orig 2025-12-04T08:53:58.7720649Z * [new branch] gh/rtimpe/34/base -> origin/gh/rtimpe/34/base 2025-12-04T08:53:58.7720716Z * [new branch] gh/rtimpe/34/head -> origin/gh/rtimpe/34/head 2025-12-04T08:53:58.7720783Z * [new branch] gh/rtimpe/34/orig -> origin/gh/rtimpe/34/orig 2025-12-04T08:53:58.7720850Z * [new branch] gh/rtimpe/35/base -> origin/gh/rtimpe/35/base 2025-12-04T08:53:58.7720919Z * [new branch] gh/rtimpe/35/head -> origin/gh/rtimpe/35/head 2025-12-04T08:53:58.7720986Z * [new branch] gh/rtimpe/35/orig -> origin/gh/rtimpe/35/orig 2025-12-04T08:53:58.7721052Z * [new branch] gh/rtimpe/4/base -> origin/gh/rtimpe/4/base 2025-12-04T08:53:58.7721121Z * [new branch] gh/rtimpe/4/head -> origin/gh/rtimpe/4/head 2025-12-04T08:53:58.7721204Z * [new branch] gh/ruisizhang123/1/base -> origin/gh/ruisizhang123/1/base 2025-12-04T08:53:58.7721284Z * [new branch] gh/ruisizhang123/1/head -> origin/gh/ruisizhang123/1/head 2025-12-04T08:53:58.7721364Z * [new branch] gh/ruisizhang123/1/orig -> origin/gh/ruisizhang123/1/orig 2025-12-04T08:53:58.7721441Z * [new branch] gh/ruisizhang123/4/base -> origin/gh/ruisizhang123/4/base 2025-12-04T08:53:58.7721517Z * [new branch] gh/ruisizhang123/4/head -> origin/gh/ruisizhang123/4/head 2025-12-04T08:53:58.7721616Z * [new branch] gh/ruisizhang123/4/orig -> origin/gh/ruisizhang123/4/orig 2025-12-04T08:53:58.7721691Z * [new branch] gh/ruisizhang123/5/base -> origin/gh/ruisizhang123/5/base 2025-12-04T08:53:58.7721767Z * [new branch] gh/ruisizhang123/5/head -> origin/gh/ruisizhang123/5/head 2025-12-04T08:53:58.7721844Z * [new branch] gh/ruisizhang123/5/orig -> origin/gh/ruisizhang123/5/orig 2025-12-04T08:53:58.7721921Z * [new branch] gh/ruisizhang123/6/base -> origin/gh/ruisizhang123/6/base 2025-12-04T08:53:58.7721997Z * [new branch] gh/ruisizhang123/6/head -> origin/gh/ruisizhang123/6/head 2025-12-04T08:53:58.7722075Z * [new branch] gh/ruisizhang123/6/orig -> origin/gh/ruisizhang123/6/orig 2025-12-04T08:53:58.7722152Z * [new branch] gh/ruisizhang123/7/base -> origin/gh/ruisizhang123/7/base 2025-12-04T08:53:58.7722229Z * [new branch] gh/ruisizhang123/7/head -> origin/gh/ruisizhang123/7/head 2025-12-04T08:53:58.7722309Z * [new branch] gh/ruisizhang123/7/orig -> origin/gh/ruisizhang123/7/orig 2025-12-04T08:53:58.7722384Z * [new branch] gh/ruisizhang123/8/base -> origin/gh/ruisizhang123/8/base 2025-12-04T08:53:58.7722463Z * [new branch] gh/ruisizhang123/8/head -> origin/gh/ruisizhang123/8/head 2025-12-04T08:53:58.7722563Z * [new branch] gh/ruisizhang123/8/orig -> origin/gh/ruisizhang123/8/orig 2025-12-04T08:53:58.7722639Z * [new branch] gh/ruisizhang123/9/base -> origin/gh/ruisizhang123/9/base 2025-12-04T08:53:58.7722717Z * [new branch] gh/ruisizhang123/9/head -> origin/gh/ruisizhang123/9/head 2025-12-04T08:53:58.7722792Z * [new branch] gh/ruisizhang123/9/orig -> origin/gh/ruisizhang123/9/orig 2025-12-04T08:53:58.7722870Z * [new branch] gh/seemethere/52/base -> origin/gh/seemethere/52/base 2025-12-04T08:53:58.7722948Z * [new branch] gh/seemethere/52/head -> origin/gh/seemethere/52/head 2025-12-04T08:53:58.7723024Z * [new branch] gh/seemethere/52/orig -> origin/gh/seemethere/52/orig 2025-12-04T08:53:58.7723098Z * [new branch] gh/seemethere/53/base -> origin/gh/seemethere/53/base 2025-12-04T08:53:58.7723173Z * [new branch] gh/seemethere/53/head -> origin/gh/seemethere/53/head 2025-12-04T08:53:58.7723247Z * [new branch] gh/seemethere/53/orig -> origin/gh/seemethere/53/orig 2025-12-04T08:53:58.7723320Z * [new branch] gh/seemethere/54/base -> origin/gh/seemethere/54/base 2025-12-04T08:53:58.7723397Z * [new branch] gh/seemethere/54/head -> origin/gh/seemethere/54/head 2025-12-04T08:53:58.7723471Z * [new branch] gh/seemethere/54/orig -> origin/gh/seemethere/54/orig 2025-12-04T08:53:58.7723545Z * [new branch] gh/seemethere/55/base -> origin/gh/seemethere/55/base 2025-12-04T08:53:58.7723621Z * [new branch] gh/seemethere/55/head -> origin/gh/seemethere/55/head 2025-12-04T08:53:58.7723694Z * [new branch] gh/seemethere/55/orig -> origin/gh/seemethere/55/orig 2025-12-04T08:53:58.7723767Z * [new branch] gh/seemethere/59/base -> origin/gh/seemethere/59/base 2025-12-04T08:53:58.7723844Z * [new branch] gh/seemethere/59/head -> origin/gh/seemethere/59/head 2025-12-04T08:53:58.7723918Z * [new branch] gh/seemethere/59/orig -> origin/gh/seemethere/59/orig 2025-12-04T08:53:58.7723992Z * [new branch] gh/seemethere/62/base -> origin/gh/seemethere/62/base 2025-12-04T08:53:58.7724065Z * [new branch] gh/seemethere/62/head -> origin/gh/seemethere/62/head 2025-12-04T08:53:58.7724140Z * [new branch] gh/seemethere/62/orig -> origin/gh/seemethere/62/orig 2025-12-04T08:53:58.7724215Z * [new branch] gh/seemethere/63/base -> origin/gh/seemethere/63/base 2025-12-04T08:53:58.7724319Z * [new branch] gh/seemethere/63/head -> origin/gh/seemethere/63/head 2025-12-04T08:53:58.7724392Z * [new branch] gh/seemethere/63/orig -> origin/gh/seemethere/63/orig 2025-12-04T08:53:58.7724467Z * [new branch] gh/seemethere/71/base -> origin/gh/seemethere/71/base 2025-12-04T08:53:58.7724541Z * [new branch] gh/seemethere/71/head -> origin/gh/seemethere/71/head 2025-12-04T08:53:58.7724615Z * [new branch] gh/seemethere/71/orig -> origin/gh/seemethere/71/orig 2025-12-04T08:53:58.7724690Z * [new branch] gh/seemethere/72/base -> origin/gh/seemethere/72/base 2025-12-04T08:53:58.7724764Z * [new branch] gh/seemethere/72/head -> origin/gh/seemethere/72/head 2025-12-04T08:53:58.7724838Z * [new branch] gh/seemethere/72/orig -> origin/gh/seemethere/72/orig 2025-12-04T08:53:58.7724912Z * [new branch] gh/seemethere/73/base -> origin/gh/seemethere/73/base 2025-12-04T08:53:58.7724988Z * [new branch] gh/seemethere/73/head -> origin/gh/seemethere/73/head 2025-12-04T08:53:58.7725061Z * [new branch] gh/seemethere/73/orig -> origin/gh/seemethere/73/orig 2025-12-04T08:53:58.7725136Z * [new branch] gh/seemethere/74/base -> origin/gh/seemethere/74/base 2025-12-04T08:53:58.7725247Z * [new branch] gh/seemethere/74/head -> origin/gh/seemethere/74/head 2025-12-04T08:53:58.7725321Z * [new branch] gh/seemethere/74/orig -> origin/gh/seemethere/74/orig 2025-12-04T08:53:58.7725395Z * [new branch] gh/seemethere/75/base -> origin/gh/seemethere/75/base 2025-12-04T08:53:58.7725468Z * [new branch] gh/seemethere/75/head -> origin/gh/seemethere/75/head 2025-12-04T08:53:58.7725543Z * [new branch] gh/seemethere/75/orig -> origin/gh/seemethere/75/orig 2025-12-04T08:53:58.7725618Z * [new branch] gh/seemethere/76/base -> origin/gh/seemethere/76/base 2025-12-04T08:53:58.7725691Z * [new branch] gh/seemethere/76/head -> origin/gh/seemethere/76/head 2025-12-04T08:53:58.7725764Z * [new branch] gh/seemethere/76/orig -> origin/gh/seemethere/76/orig 2025-12-04T08:53:58.7725843Z * [new branch] gh/shunting314/145/base -> origin/gh/shunting314/145/base 2025-12-04T08:53:58.7725923Z * [new branch] gh/shunting314/145/head -> origin/gh/shunting314/145/head 2025-12-04T08:53:58.7726001Z * [new branch] gh/shunting314/145/orig -> origin/gh/shunting314/145/orig 2025-12-04T08:53:58.7726076Z * [new branch] gh/shunting314/176/base -> origin/gh/shunting314/176/base 2025-12-04T08:53:58.7726151Z * [new branch] gh/shunting314/176/head -> origin/gh/shunting314/176/head 2025-12-04T08:53:58.7726228Z * [new branch] gh/shunting314/176/orig -> origin/gh/shunting314/176/orig 2025-12-04T08:53:58.7726303Z * [new branch] gh/shunting314/249/base -> origin/gh/shunting314/249/base 2025-12-04T08:53:58.7726377Z * [new branch] gh/shunting314/249/head -> origin/gh/shunting314/249/head 2025-12-04T08:53:58.7726454Z * [new branch] gh/shunting314/249/orig -> origin/gh/shunting314/249/orig 2025-12-04T08:53:58.7726529Z * [new branch] gh/shunting314/253/base -> origin/gh/shunting314/253/base 2025-12-04T08:53:58.7726604Z * [new branch] gh/shunting314/253/head -> origin/gh/shunting314/253/head 2025-12-04T08:53:58.7726678Z * [new branch] gh/shunting314/253/orig -> origin/gh/shunting314/253/orig 2025-12-04T08:53:58.7726799Z * [new branch] gh/shunting314/256/base -> origin/gh/shunting314/256/base 2025-12-04T08:53:58.7726875Z * [new branch] gh/shunting314/256/head -> origin/gh/shunting314/256/head 2025-12-04T08:53:58.7726951Z * [new branch] gh/shunting314/256/orig -> origin/gh/shunting314/256/orig 2025-12-04T08:53:58.7727064Z * [new branch] gh/shunting314/257/base -> origin/gh/shunting314/257/base 2025-12-04T08:53:58.7727139Z * [new branch] gh/shunting314/257/head -> origin/gh/shunting314/257/head 2025-12-04T08:53:58.7727215Z * [new branch] gh/shunting314/257/orig -> origin/gh/shunting314/257/orig 2025-12-04T08:53:58.7727291Z * [new branch] gh/shunting314/258/base -> origin/gh/shunting314/258/base 2025-12-04T08:53:58.7727367Z * [new branch] gh/shunting314/258/head -> origin/gh/shunting314/258/head 2025-12-04T08:53:58.7727442Z * [new branch] gh/shunting314/258/orig -> origin/gh/shunting314/258/orig 2025-12-04T08:53:58.7727516Z * [new branch] gh/shunting314/259/base -> origin/gh/shunting314/259/base 2025-12-04T08:53:58.7727593Z * [new branch] gh/shunting314/259/head -> origin/gh/shunting314/259/head 2025-12-04T08:53:58.7727669Z * [new branch] gh/shunting314/259/orig -> origin/gh/shunting314/259/orig 2025-12-04T08:53:58.7727745Z * [new branch] gh/shunting314/260/base -> origin/gh/shunting314/260/base 2025-12-04T08:53:58.7727820Z * [new branch] gh/shunting314/260/head -> origin/gh/shunting314/260/head 2025-12-04T08:53:58.7727894Z * [new branch] gh/shunting314/260/orig -> origin/gh/shunting314/260/orig 2025-12-04T08:53:58.7728005Z * [new branch] gh/shunting314/261/base -> origin/gh/shunting314/261/base 2025-12-04T08:53:58.7728178Z * [new branch] gh/shunting314/261/head -> origin/gh/shunting314/261/head 2025-12-04T08:53:58.7728251Z * [new branch] gh/shunting314/261/orig -> origin/gh/shunting314/261/orig 2025-12-04T08:53:58.7728325Z * [new branch] gh/shunting314/262/base -> origin/gh/shunting314/262/base 2025-12-04T08:53:58.7728400Z * [new branch] gh/shunting314/262/head -> origin/gh/shunting314/262/head 2025-12-04T08:53:58.7728476Z * [new branch] gh/shunting314/262/orig -> origin/gh/shunting314/262/orig 2025-12-04T08:53:58.7728551Z * [new branch] gh/shunting314/263/base -> origin/gh/shunting314/263/base 2025-12-04T08:53:58.7728627Z * [new branch] gh/shunting314/263/head -> origin/gh/shunting314/263/head 2025-12-04T08:53:58.7728703Z * [new branch] gh/shunting314/263/orig -> origin/gh/shunting314/263/orig 2025-12-04T08:53:58.7728778Z * [new branch] gh/shunting314/264/base -> origin/gh/shunting314/264/base 2025-12-04T08:53:58.7728853Z * [new branch] gh/shunting314/264/head -> origin/gh/shunting314/264/head 2025-12-04T08:53:58.7728927Z * [new branch] gh/shunting314/264/orig -> origin/gh/shunting314/264/orig 2025-12-04T08:53:58.7729002Z * [new branch] gh/shunting314/265/base -> origin/gh/shunting314/265/base 2025-12-04T08:53:58.7729075Z * [new branch] gh/shunting314/265/head -> origin/gh/shunting314/265/head 2025-12-04T08:53:58.7729152Z * [new branch] gh/shunting314/265/orig -> origin/gh/shunting314/265/orig 2025-12-04T08:53:58.7729228Z * [new branch] gh/shunting314/266/base -> origin/gh/shunting314/266/base 2025-12-04T08:53:58.7729302Z * [new branch] gh/shunting314/266/head -> origin/gh/shunting314/266/head 2025-12-04T08:53:58.7729377Z * [new branch] gh/shunting314/266/orig -> origin/gh/shunting314/266/orig 2025-12-04T08:53:58.7729454Z * [new branch] gh/shunting314/267/base -> origin/gh/shunting314/267/base 2025-12-04T08:53:58.7729528Z * [new branch] gh/shunting314/267/head -> origin/gh/shunting314/267/head 2025-12-04T08:53:58.7729602Z * [new branch] gh/shunting314/267/orig -> origin/gh/shunting314/267/orig 2025-12-04T08:53:58.7729676Z * [new branch] gh/shunting314/268/base -> origin/gh/shunting314/268/base 2025-12-04T08:53:58.7729750Z * [new branch] gh/shunting314/268/head -> origin/gh/shunting314/268/head 2025-12-04T08:53:58.7729859Z * [new branch] gh/shunting314/268/orig -> origin/gh/shunting314/268/orig 2025-12-04T08:53:58.7729935Z * [new branch] gh/shunting314/269/base -> origin/gh/shunting314/269/base 2025-12-04T08:53:58.7730009Z * [new branch] gh/shunting314/269/head -> origin/gh/shunting314/269/head 2025-12-04T08:53:58.7730084Z * [new branch] gh/shunting314/269/orig -> origin/gh/shunting314/269/orig 2025-12-04T08:53:58.7730161Z * [new branch] gh/silverguo/1/base -> origin/gh/silverguo/1/base 2025-12-04T08:53:58.7730233Z * [new branch] gh/silverguo/1/head -> origin/gh/silverguo/1/head 2025-12-04T08:53:58.7730306Z * [new branch] gh/silverguo/2/base -> origin/gh/silverguo/2/base 2025-12-04T08:53:58.7730379Z * [new branch] gh/silverguo/2/head -> origin/gh/silverguo/2/head 2025-12-04T08:53:58.7730453Z * [new branch] gh/silverguo/3/base -> origin/gh/silverguo/3/base 2025-12-04T08:53:58.7730524Z * [new branch] gh/silverguo/3/head -> origin/gh/silverguo/3/head 2025-12-04T08:53:58.7730596Z * [new branch] gh/silverguo/4/base -> origin/gh/silverguo/4/base 2025-12-04T08:53:58.7730668Z * [new branch] gh/silverguo/4/head -> origin/gh/silverguo/4/head 2025-12-04T08:53:58.7730766Z * [new branch] gh/slayton58/39/base -> origin/gh/slayton58/39/base 2025-12-04T08:53:58.7730839Z * [new branch] gh/slayton58/39/head -> origin/gh/slayton58/39/head 2025-12-04T08:53:58.7730911Z * [new branch] gh/slayton58/39/orig -> origin/gh/slayton58/39/orig 2025-12-04T08:53:58.7730986Z * [new branch] gh/slayton58/42/base -> origin/gh/slayton58/42/base 2025-12-04T08:53:58.7731057Z * [new branch] gh/slayton58/42/head -> origin/gh/slayton58/42/head 2025-12-04T08:53:58.7731131Z * [new branch] gh/slayton58/42/orig -> origin/gh/slayton58/42/orig 2025-12-04T08:53:58.7731203Z * [new branch] gh/slayton58/43/base -> origin/gh/slayton58/43/base 2025-12-04T08:53:58.7731274Z * [new branch] gh/slayton58/43/head -> origin/gh/slayton58/43/head 2025-12-04T08:53:58.7731344Z * [new branch] gh/slayton58/43/orig -> origin/gh/slayton58/43/orig 2025-12-04T08:53:58.7731417Z * [new branch] gh/slayton58/44/base -> origin/gh/slayton58/44/base 2025-12-04T08:53:58.7731488Z * [new branch] gh/slayton58/44/head -> origin/gh/slayton58/44/head 2025-12-04T08:53:58.7731560Z * [new branch] gh/slayton58/44/orig -> origin/gh/slayton58/44/orig 2025-12-04T08:53:58.7731630Z * [new branch] gh/slayton58/45/base -> origin/gh/slayton58/45/base 2025-12-04T08:53:58.7731700Z * [new branch] gh/slayton58/45/head -> origin/gh/slayton58/45/head 2025-12-04T08:53:58.7731772Z * [new branch] gh/slayton58/45/orig -> origin/gh/slayton58/45/orig 2025-12-04T08:53:58.7731845Z * [new branch] gh/slayton58/46/base -> origin/gh/slayton58/46/base 2025-12-04T08:53:58.7731916Z * [new branch] gh/slayton58/46/head -> origin/gh/slayton58/46/head 2025-12-04T08:53:58.7731987Z * [new branch] gh/slayton58/46/orig -> origin/gh/slayton58/46/orig 2025-12-04T08:53:58.7732058Z * [new branch] gh/slayton58/6/base -> origin/gh/slayton58/6/base 2025-12-04T08:53:58.7732128Z * [new branch] gh/slayton58/6/head -> origin/gh/slayton58/6/head 2025-12-04T08:53:58.7732198Z * [new branch] gh/slayton58/7/base -> origin/gh/slayton58/7/base 2025-12-04T08:53:58.7732267Z * [new branch] gh/slayton58/7/head -> origin/gh/slayton58/7/head 2025-12-04T08:53:58.7732341Z * [new branch] gh/soulitzer/269/base -> origin/gh/soulitzer/269/base 2025-12-04T08:53:58.7732445Z * [new branch] gh/soulitzer/269/head -> origin/gh/soulitzer/269/head 2025-12-04T08:53:58.7732519Z * [new branch] gh/soulitzer/269/orig -> origin/gh/soulitzer/269/orig 2025-12-04T08:53:58.7732592Z * [new branch] gh/soulitzer/276/base -> origin/gh/soulitzer/276/base 2025-12-04T08:53:58.7732666Z * [new branch] gh/soulitzer/276/head -> origin/gh/soulitzer/276/head 2025-12-04T08:53:58.7732739Z * [new branch] gh/soulitzer/276/orig -> origin/gh/soulitzer/276/orig 2025-12-04T08:53:58.7732811Z * [new branch] gh/soulitzer/287/base -> origin/gh/soulitzer/287/base 2025-12-04T08:53:58.7732885Z * [new branch] gh/soulitzer/287/head -> origin/gh/soulitzer/287/head 2025-12-04T08:53:58.7732959Z * [new branch] gh/soulitzer/287/orig -> origin/gh/soulitzer/287/orig 2025-12-04T08:53:58.7733033Z * [new branch] gh/soulitzer/296/base -> origin/gh/soulitzer/296/base 2025-12-04T08:53:58.7733107Z * [new branch] gh/soulitzer/296/head -> origin/gh/soulitzer/296/head 2025-12-04T08:53:58.7733179Z * [new branch] gh/soulitzer/296/orig -> origin/gh/soulitzer/296/orig 2025-12-04T08:53:58.7733251Z * [new branch] gh/soulitzer/299/base -> origin/gh/soulitzer/299/base 2025-12-04T08:53:58.7733349Z * [new branch] gh/soulitzer/299/head -> origin/gh/soulitzer/299/head 2025-12-04T08:53:58.7733423Z * [new branch] gh/soulitzer/299/orig -> origin/gh/soulitzer/299/orig 2025-12-04T08:53:58.7733497Z * [new branch] gh/soulitzer/300/base -> origin/gh/soulitzer/300/base 2025-12-04T08:53:58.7733569Z * [new branch] gh/soulitzer/300/head -> origin/gh/soulitzer/300/head 2025-12-04T08:53:58.7733642Z * [new branch] gh/soulitzer/300/orig -> origin/gh/soulitzer/300/orig 2025-12-04T08:53:58.7733719Z * [new branch] gh/soulitzer/301/base -> origin/gh/soulitzer/301/base 2025-12-04T08:53:58.7733793Z * [new branch] gh/soulitzer/301/head -> origin/gh/soulitzer/301/head 2025-12-04T08:53:58.7733865Z * [new branch] gh/soulitzer/301/orig -> origin/gh/soulitzer/301/orig 2025-12-04T08:53:58.7733939Z * [new branch] gh/soulitzer/313/base -> origin/gh/soulitzer/313/base 2025-12-04T08:53:58.7734013Z * [new branch] gh/soulitzer/313/head -> origin/gh/soulitzer/313/head 2025-12-04T08:53:58.7734085Z * [new branch] gh/soulitzer/313/orig -> origin/gh/soulitzer/313/orig 2025-12-04T08:53:58.7734159Z * [new branch] gh/soulitzer/319/base -> origin/gh/soulitzer/319/base 2025-12-04T08:53:58.7734231Z * [new branch] gh/soulitzer/319/head -> origin/gh/soulitzer/319/head 2025-12-04T08:53:58.7734304Z * [new branch] gh/soulitzer/319/orig -> origin/gh/soulitzer/319/orig 2025-12-04T08:53:58.7734379Z * [new branch] gh/soulitzer/320/base -> origin/gh/soulitzer/320/base 2025-12-04T08:53:58.7734454Z * [new branch] gh/soulitzer/320/head -> origin/gh/soulitzer/320/head 2025-12-04T08:53:58.7734526Z * [new branch] gh/soulitzer/320/orig -> origin/gh/soulitzer/320/orig 2025-12-04T08:53:58.7734600Z * [new branch] gh/soulitzer/336/base -> origin/gh/soulitzer/336/base 2025-12-04T08:53:58.7734673Z * [new branch] gh/soulitzer/336/head -> origin/gh/soulitzer/336/head 2025-12-04T08:53:58.7734746Z * [new branch] gh/soulitzer/336/orig -> origin/gh/soulitzer/336/orig 2025-12-04T08:53:58.7734820Z * [new branch] gh/soulitzer/347/base -> origin/gh/soulitzer/347/base 2025-12-04T08:53:58.7734892Z * [new branch] gh/soulitzer/347/head -> origin/gh/soulitzer/347/head 2025-12-04T08:53:58.7734964Z * [new branch] gh/soulitzer/347/orig -> origin/gh/soulitzer/347/orig 2025-12-04T08:53:58.7735077Z * [new branch] gh/soulitzer/349/base -> origin/gh/soulitzer/349/base 2025-12-04T08:53:58.7735152Z * [new branch] gh/soulitzer/349/head -> origin/gh/soulitzer/349/head 2025-12-04T08:53:58.7735226Z * [new branch] gh/soulitzer/349/orig -> origin/gh/soulitzer/349/orig 2025-12-04T08:53:58.7735298Z * [new branch] gh/soulitzer/350/base -> origin/gh/soulitzer/350/base 2025-12-04T08:53:58.7735373Z * [new branch] gh/soulitzer/350/head -> origin/gh/soulitzer/350/head 2025-12-04T08:53:58.7735446Z * [new branch] gh/soulitzer/350/orig -> origin/gh/soulitzer/350/orig 2025-12-04T08:53:58.7735518Z * [new branch] gh/soulitzer/351/base -> origin/gh/soulitzer/351/base 2025-12-04T08:53:58.7735589Z * [new branch] gh/soulitzer/351/head -> origin/gh/soulitzer/351/head 2025-12-04T08:53:58.7735662Z * [new branch] gh/soulitzer/351/orig -> origin/gh/soulitzer/351/orig 2025-12-04T08:53:58.7735736Z * [new branch] gh/soulitzer/353/base -> origin/gh/soulitzer/353/base 2025-12-04T08:53:58.7735809Z * [new branch] gh/soulitzer/353/head -> origin/gh/soulitzer/353/head 2025-12-04T08:53:58.7735884Z * [new branch] gh/soulitzer/353/orig -> origin/gh/soulitzer/353/orig 2025-12-04T08:53:58.7735956Z * [new branch] gh/soulitzer/358/base -> origin/gh/soulitzer/358/base 2025-12-04T08:53:58.7736057Z * [new branch] gh/soulitzer/358/head -> origin/gh/soulitzer/358/head 2025-12-04T08:53:58.7736134Z * [new branch] gh/soulitzer/358/orig -> origin/gh/soulitzer/358/orig 2025-12-04T08:53:58.7736208Z * [new branch] gh/soulitzer/359/base -> origin/gh/soulitzer/359/base 2025-12-04T08:53:58.7736280Z * [new branch] gh/soulitzer/359/head -> origin/gh/soulitzer/359/head 2025-12-04T08:53:58.7736352Z * [new branch] gh/soulitzer/359/orig -> origin/gh/soulitzer/359/orig 2025-12-04T08:53:58.7736426Z * [new branch] gh/soulitzer/374/base -> origin/gh/soulitzer/374/base 2025-12-04T08:53:58.7736499Z * [new branch] gh/soulitzer/374/head -> origin/gh/soulitzer/374/head 2025-12-04T08:53:58.7736573Z * [new branch] gh/soulitzer/374/orig -> origin/gh/soulitzer/374/orig 2025-12-04T08:53:58.7736647Z * [new branch] gh/soulitzer/375/base -> origin/gh/soulitzer/375/base 2025-12-04T08:53:58.7736720Z * [new branch] gh/soulitzer/375/head -> origin/gh/soulitzer/375/head 2025-12-04T08:53:58.7736833Z * [new branch] gh/soulitzer/375/orig -> origin/gh/soulitzer/375/orig 2025-12-04T08:53:58.7736907Z * [new branch] gh/soulitzer/380/base -> origin/gh/soulitzer/380/base 2025-12-04T08:53:58.7736981Z * [new branch] gh/soulitzer/380/head -> origin/gh/soulitzer/380/head 2025-12-04T08:53:58.7737053Z * [new branch] gh/soulitzer/380/orig -> origin/gh/soulitzer/380/orig 2025-12-04T08:53:58.7737126Z * [new branch] gh/soulitzer/385/base -> origin/gh/soulitzer/385/base 2025-12-04T08:53:58.7737200Z * [new branch] gh/soulitzer/385/head -> origin/gh/soulitzer/385/head 2025-12-04T08:53:58.7737272Z * [new branch] gh/soulitzer/385/orig -> origin/gh/soulitzer/385/orig 2025-12-04T08:53:58.7737346Z * [new branch] gh/soulitzer/386/base -> origin/gh/soulitzer/386/base 2025-12-04T08:53:58.7737420Z * [new branch] gh/soulitzer/386/head -> origin/gh/soulitzer/386/head 2025-12-04T08:53:58.7737492Z * [new branch] gh/soulitzer/386/orig -> origin/gh/soulitzer/386/orig 2025-12-04T08:53:58.7737564Z * [new branch] gh/soulitzer/387/base -> origin/gh/soulitzer/387/base 2025-12-04T08:53:58.7737637Z * [new branch] gh/soulitzer/387/head -> origin/gh/soulitzer/387/head 2025-12-04T08:53:58.7737709Z * [new branch] gh/soulitzer/387/orig -> origin/gh/soulitzer/387/orig 2025-12-04T08:53:58.7737825Z * [new branch] gh/soulitzer/388/base -> origin/gh/soulitzer/388/base 2025-12-04T08:53:58.7737902Z * [new branch] gh/soulitzer/388/head -> origin/gh/soulitzer/388/head 2025-12-04T08:53:58.7737974Z * [new branch] gh/soulitzer/388/orig -> origin/gh/soulitzer/388/orig 2025-12-04T08:53:58.7738047Z * [new branch] gh/soulitzer/389/base -> origin/gh/soulitzer/389/base 2025-12-04T08:53:58.7738121Z * [new branch] gh/soulitzer/389/head -> origin/gh/soulitzer/389/head 2025-12-04T08:53:58.7738194Z * [new branch] gh/soulitzer/389/orig -> origin/gh/soulitzer/389/orig 2025-12-04T08:53:58.7738268Z * [new branch] gh/soulitzer/390/base -> origin/gh/soulitzer/390/base 2025-12-04T08:53:58.7738340Z * [new branch] gh/soulitzer/390/head -> origin/gh/soulitzer/390/head 2025-12-04T08:53:58.7738412Z * [new branch] gh/soulitzer/390/orig -> origin/gh/soulitzer/390/orig 2025-12-04T08:53:58.7738487Z * [new branch] gh/soulitzer/391/base -> origin/gh/soulitzer/391/base 2025-12-04T08:53:58.7738559Z * [new branch] gh/soulitzer/391/head -> origin/gh/soulitzer/391/head 2025-12-04T08:53:58.7738633Z * [new branch] gh/soulitzer/391/orig -> origin/gh/soulitzer/391/orig 2025-12-04T08:53:58.7738742Z * [new branch] gh/soulitzer/392/base -> origin/gh/soulitzer/392/base 2025-12-04T08:53:58.7738815Z * [new branch] gh/soulitzer/392/head -> origin/gh/soulitzer/392/head 2025-12-04T08:53:58.7738887Z * [new branch] gh/soulitzer/392/orig -> origin/gh/soulitzer/392/orig 2025-12-04T08:53:58.7738960Z * [new branch] gh/swolchok/728/next -> origin/gh/swolchok/728/next 2025-12-04T08:53:58.7739031Z * [new branch] gh/swolchok/819/base -> origin/gh/swolchok/819/base 2025-12-04T08:53:58.7739103Z * [new branch] gh/swolchok/819/head -> origin/gh/swolchok/819/head 2025-12-04T08:53:58.7739176Z * [new branch] gh/swolchok/819/orig -> origin/gh/swolchok/819/orig 2025-12-04T08:53:58.7739246Z * [new branch] gh/swolchok/824/base -> origin/gh/swolchok/824/base 2025-12-04T08:53:58.7739316Z * [new branch] gh/swolchok/824/head -> origin/gh/swolchok/824/head 2025-12-04T08:53:58.7739391Z * [new branch] gh/swolchok/824/orig -> origin/gh/swolchok/824/orig 2025-12-04T08:53:58.7739463Z * [new branch] gh/swolchok/829/base -> origin/gh/swolchok/829/base 2025-12-04T08:53:58.7739535Z * [new branch] gh/swolchok/829/head -> origin/gh/swolchok/829/head 2025-12-04T08:53:58.7739609Z * [new branch] gh/swolchok/829/orig -> origin/gh/swolchok/829/orig 2025-12-04T08:53:58.7739680Z * [new branch] gh/swolchok/839/base -> origin/gh/swolchok/839/base 2025-12-04T08:53:58.7739754Z * [new branch] gh/swolchok/839/head -> origin/gh/swolchok/839/head 2025-12-04T08:53:58.7739826Z * [new branch] gh/swolchok/839/orig -> origin/gh/swolchok/839/orig 2025-12-04T08:53:58.7739896Z * [new branch] gh/swolchok/841/base -> origin/gh/swolchok/841/base 2025-12-04T08:53:58.7739966Z * [new branch] gh/swolchok/841/head -> origin/gh/swolchok/841/head 2025-12-04T08:53:58.7740039Z * [new branch] gh/swolchok/841/orig -> origin/gh/swolchok/841/orig 2025-12-04T08:53:58.7740111Z * [new branch] gh/swolchok/842/base -> origin/gh/swolchok/842/base 2025-12-04T08:53:58.7740185Z * [new branch] gh/swolchok/842/head -> origin/gh/swolchok/842/head 2025-12-04T08:53:58.7740258Z * [new branch] gh/swolchok/842/orig -> origin/gh/swolchok/842/orig 2025-12-04T08:53:58.7740329Z * [new branch] gh/swolchok/845/base -> origin/gh/swolchok/845/base 2025-12-04T08:53:58.7740426Z * [new branch] gh/swolchok/845/head -> origin/gh/swolchok/845/head 2025-12-04T08:53:58.7740498Z * [new branch] gh/swolchok/845/orig -> origin/gh/swolchok/845/orig 2025-12-04T08:53:58.7740568Z * [new branch] gh/swolchok/848/base -> origin/gh/swolchok/848/base 2025-12-04T08:53:58.7740640Z * [new branch] gh/swolchok/848/head -> origin/gh/swolchok/848/head 2025-12-04T08:53:58.7740712Z * [new branch] gh/swolchok/848/orig -> origin/gh/swolchok/848/orig 2025-12-04T08:53:58.7740784Z * [new branch] gh/swolchok/856/base -> origin/gh/swolchok/856/base 2025-12-04T08:53:58.7740857Z * [new branch] gh/swolchok/856/head -> origin/gh/swolchok/856/head 2025-12-04T08:53:58.7740927Z * [new branch] gh/swolchok/856/orig -> origin/gh/swolchok/856/orig 2025-12-04T08:53:58.7740998Z * [new branch] gh/swolchok/860/base -> origin/gh/swolchok/860/base 2025-12-04T08:53:58.7741073Z * [new branch] gh/swolchok/860/head -> origin/gh/swolchok/860/head 2025-12-04T08:53:58.7741143Z * [new branch] gh/swolchok/860/orig -> origin/gh/swolchok/860/orig 2025-12-04T08:53:58.7741213Z * [new branch] gh/swolchok/861/base -> origin/gh/swolchok/861/base 2025-12-04T08:53:58.7741286Z * [new branch] gh/swolchok/861/head -> origin/gh/swolchok/861/head 2025-12-04T08:53:58.7741386Z * [new branch] gh/swolchok/861/orig -> origin/gh/swolchok/861/orig 2025-12-04T08:53:58.7741458Z * [new branch] gh/swolchok/862/base -> origin/gh/swolchok/862/base 2025-12-04T08:53:58.7741528Z * [new branch] gh/swolchok/862/head -> origin/gh/swolchok/862/head 2025-12-04T08:53:58.7741597Z * [new branch] gh/swolchok/862/orig -> origin/gh/swolchok/862/orig 2025-12-04T08:53:58.7741671Z * [new branch] gh/swolchok/863/base -> origin/gh/swolchok/863/base 2025-12-04T08:53:58.7741744Z * [new branch] gh/swolchok/863/head -> origin/gh/swolchok/863/head 2025-12-04T08:53:58.7741815Z * [new branch] gh/swolchok/863/orig -> origin/gh/swolchok/863/orig 2025-12-04T08:53:58.7741887Z * [new branch] gh/swolchok/864/base -> origin/gh/swolchok/864/base 2025-12-04T08:53:58.7741958Z * [new branch] gh/swolchok/864/head -> origin/gh/swolchok/864/head 2025-12-04T08:53:58.7742028Z * [new branch] gh/swolchok/864/orig -> origin/gh/swolchok/864/orig 2025-12-04T08:53:58.7742099Z * [new branch] gh/swolchok/865/base -> origin/gh/swolchok/865/base 2025-12-04T08:53:58.7742169Z * [new branch] gh/swolchok/865/head -> origin/gh/swolchok/865/head 2025-12-04T08:53:58.7742239Z * [new branch] gh/swolchok/865/orig -> origin/gh/swolchok/865/orig 2025-12-04T08:53:58.7742311Z * [new branch] gh/swolchok/866/base -> origin/gh/swolchok/866/base 2025-12-04T08:53:58.7742384Z * [new branch] gh/swolchok/866/head -> origin/gh/swolchok/866/head 2025-12-04T08:53:58.7742455Z * [new branch] gh/swolchok/866/orig -> origin/gh/swolchok/866/orig 2025-12-04T08:53:58.7742528Z * [new branch] gh/swolchok/867/base -> origin/gh/swolchok/867/base 2025-12-04T08:53:58.7742599Z * [new branch] gh/swolchok/867/head -> origin/gh/swolchok/867/head 2025-12-04T08:53:58.7742671Z * [new branch] gh/swolchok/867/orig -> origin/gh/swolchok/867/orig 2025-12-04T08:53:58.7742742Z * [new branch] gh/swolchok/868/base -> origin/gh/swolchok/868/base 2025-12-04T08:53:58.7742814Z * [new branch] gh/swolchok/868/head -> origin/gh/swolchok/868/head 2025-12-04T08:53:58.7742886Z * [new branch] gh/swolchok/868/orig -> origin/gh/swolchok/868/orig 2025-12-04T08:53:58.7742957Z * [new branch] gh/swolchok/869/base -> origin/gh/swolchok/869/base 2025-12-04T08:53:58.7743050Z * [new branch] gh/swolchok/869/head -> origin/gh/swolchok/869/head 2025-12-04T08:53:58.7743124Z * [new branch] gh/swolchok/869/orig -> origin/gh/swolchok/869/orig 2025-12-04T08:53:58.7743195Z * [new branch] gh/swolchok/870/base -> origin/gh/swolchok/870/base 2025-12-04T08:53:58.7743267Z * [new branch] gh/swolchok/870/head -> origin/gh/swolchok/870/head 2025-12-04T08:53:58.7743341Z * [new branch] gh/swolchok/870/orig -> origin/gh/swolchok/870/orig 2025-12-04T08:53:58.7743411Z * [new branch] gh/swolchok/871/base -> origin/gh/swolchok/871/base 2025-12-04T08:53:58.7743482Z * [new branch] gh/swolchok/871/head -> origin/gh/swolchok/871/head 2025-12-04T08:53:58.7743556Z * [new branch] gh/swolchok/871/orig -> origin/gh/swolchok/871/orig 2025-12-04T08:53:58.7743631Z * [new branch] gh/teja-rao/4/base -> origin/gh/teja-rao/4/base 2025-12-04T08:53:58.7743705Z * [new branch] gh/teja-rao/4/head -> origin/gh/teja-rao/4/head 2025-12-04T08:53:58.7743776Z * [new branch] gh/teja-rao/4/orig -> origin/gh/teja-rao/4/orig 2025-12-04T08:53:58.7743850Z * [new branch] gh/tianyu-l/2/base -> origin/gh/tianyu-l/2/base 2025-12-04T08:53:58.7743947Z * [new branch] gh/tianyu-l/2/head -> origin/gh/tianyu-l/2/head 2025-12-04T08:53:58.7744021Z * [new branch] gh/tianyu-l/2/orig -> origin/gh/tianyu-l/2/orig 2025-12-04T08:53:58.7744089Z * [new branch] gh/tianyu-l/3/base -> origin/gh/tianyu-l/3/base 2025-12-04T08:53:58.7744157Z * [new branch] gh/tianyu-l/3/orig -> origin/gh/tianyu-l/3/orig 2025-12-04T08:53:58.7744229Z * [new branch] gh/tianyu-l/4/base -> origin/gh/tianyu-l/4/base 2025-12-04T08:53:58.7744299Z * [new branch] gh/tianyu-l/4/head -> origin/gh/tianyu-l/4/head 2025-12-04T08:53:58.7744370Z * [new branch] gh/tianyu-l/4/orig -> origin/gh/tianyu-l/4/orig 2025-12-04T08:53:58.7744461Z * [new branch] gh/tugsbayasgalan/10/base -> origin/gh/tugsbayasgalan/10/base 2025-12-04T08:53:58.7744548Z * [new branch] gh/tugsbayasgalan/10/head -> origin/gh/tugsbayasgalan/10/head 2025-12-04T08:53:58.7744637Z * [new branch] gh/tugsbayasgalan/10/orig -> origin/gh/tugsbayasgalan/10/orig 2025-12-04T08:53:58.7744722Z * [new branch] gh/tugsbayasgalan/13/base -> origin/gh/tugsbayasgalan/13/base 2025-12-04T08:53:58.7744805Z * [new branch] gh/tugsbayasgalan/13/head -> origin/gh/tugsbayasgalan/13/head 2025-12-04T08:53:58.7744889Z * [new branch] gh/tugsbayasgalan/13/orig -> origin/gh/tugsbayasgalan/13/orig 2025-12-04T08:53:58.7744972Z * [new branch] gh/tugsbayasgalan/17/base -> origin/gh/tugsbayasgalan/17/base 2025-12-04T08:53:58.7745058Z * [new branch] gh/tugsbayasgalan/17/head -> origin/gh/tugsbayasgalan/17/head 2025-12-04T08:53:58.7745141Z * [new branch] gh/tugsbayasgalan/17/orig -> origin/gh/tugsbayasgalan/17/orig 2025-12-04T08:53:58.7745224Z * [new branch] gh/tugsbayasgalan/2/base -> origin/gh/tugsbayasgalan/2/base 2025-12-04T08:53:58.7745307Z * [new branch] gh/tugsbayasgalan/2/head -> origin/gh/tugsbayasgalan/2/head 2025-12-04T08:53:58.7745392Z * [new branch] gh/tugsbayasgalan/2/orig -> origin/gh/tugsbayasgalan/2/orig 2025-12-04T08:53:58.7745476Z * [new branch] gh/tugsbayasgalan/28/base -> origin/gh/tugsbayasgalan/28/base 2025-12-04T08:53:58.7745560Z * [new branch] gh/tugsbayasgalan/28/head -> origin/gh/tugsbayasgalan/28/head 2025-12-04T08:53:58.7745645Z * [new branch] gh/tugsbayasgalan/28/orig -> origin/gh/tugsbayasgalan/28/orig 2025-12-04T08:53:58.7745729Z * [new branch] gh/tugsbayasgalan/32/base -> origin/gh/tugsbayasgalan/32/base 2025-12-04T08:53:58.7745848Z * [new branch] gh/tugsbayasgalan/32/head -> origin/gh/tugsbayasgalan/32/head 2025-12-04T08:53:58.7745935Z * [new branch] gh/tugsbayasgalan/32/orig -> origin/gh/tugsbayasgalan/32/orig 2025-12-04T08:53:58.7746017Z * [new branch] gh/tugsbayasgalan/35/base -> origin/gh/tugsbayasgalan/35/base 2025-12-04T08:53:58.7746102Z * [new branch] gh/tugsbayasgalan/35/head -> origin/gh/tugsbayasgalan/35/head 2025-12-04T08:53:58.7746185Z * [new branch] gh/tugsbayasgalan/35/orig -> origin/gh/tugsbayasgalan/35/orig 2025-12-04T08:53:58.7746269Z * [new branch] gh/tugsbayasgalan/36/base -> origin/gh/tugsbayasgalan/36/base 2025-12-04T08:53:58.7746353Z * [new branch] gh/tugsbayasgalan/36/head -> origin/gh/tugsbayasgalan/36/head 2025-12-04T08:53:58.7746435Z * [new branch] gh/tugsbayasgalan/36/orig -> origin/gh/tugsbayasgalan/36/orig 2025-12-04T08:53:58.7746519Z * [new branch] gh/tugsbayasgalan/37/base -> origin/gh/tugsbayasgalan/37/base 2025-12-04T08:53:58.7746604Z * [new branch] gh/tugsbayasgalan/37/head -> origin/gh/tugsbayasgalan/37/head 2025-12-04T08:53:58.7746686Z * [new branch] gh/tugsbayasgalan/37/orig -> origin/gh/tugsbayasgalan/37/orig 2025-12-04T08:53:58.7746852Z * [new branch] gh/tugsbayasgalan/43/base -> origin/gh/tugsbayasgalan/43/base 2025-12-04T08:53:58.7746938Z * [new branch] gh/tugsbayasgalan/43/head -> origin/gh/tugsbayasgalan/43/head 2025-12-04T08:53:58.7747020Z * [new branch] gh/tugsbayasgalan/43/orig -> origin/gh/tugsbayasgalan/43/orig 2025-12-04T08:53:58.7747102Z * [new branch] gh/tugsbayasgalan/48/base -> origin/gh/tugsbayasgalan/48/base 2025-12-04T08:53:58.7747185Z * [new branch] gh/tugsbayasgalan/48/head -> origin/gh/tugsbayasgalan/48/head 2025-12-04T08:53:58.7747268Z * [new branch] gh/tugsbayasgalan/48/orig -> origin/gh/tugsbayasgalan/48/orig 2025-12-04T08:53:58.7747352Z * [new branch] gh/tugsbayasgalan/51/base -> origin/gh/tugsbayasgalan/51/base 2025-12-04T08:53:58.7747435Z * [new branch] gh/tugsbayasgalan/51/head -> origin/gh/tugsbayasgalan/51/head 2025-12-04T08:53:58.7747517Z * [new branch] gh/tugsbayasgalan/51/orig -> origin/gh/tugsbayasgalan/51/orig 2025-12-04T08:53:58.7747600Z * [new branch] gh/tugsbayasgalan/52/base -> origin/gh/tugsbayasgalan/52/base 2025-12-04T08:53:58.7747684Z * [new branch] gh/tugsbayasgalan/52/head -> origin/gh/tugsbayasgalan/52/head 2025-12-04T08:53:58.7747767Z * [new branch] gh/tugsbayasgalan/52/orig -> origin/gh/tugsbayasgalan/52/orig 2025-12-04T08:53:58.7747849Z * [new branch] gh/tugsbayasgalan/53/base -> origin/gh/tugsbayasgalan/53/base 2025-12-04T08:53:58.7747931Z * [new branch] gh/tugsbayasgalan/53/head -> origin/gh/tugsbayasgalan/53/head 2025-12-04T08:53:58.7748015Z * [new branch] gh/tugsbayasgalan/53/orig -> origin/gh/tugsbayasgalan/53/orig 2025-12-04T08:53:58.7748098Z * [new branch] gh/tugsbayasgalan/55/base -> origin/gh/tugsbayasgalan/55/base 2025-12-04T08:53:58.7748181Z * [new branch] gh/tugsbayasgalan/55/head -> origin/gh/tugsbayasgalan/55/head 2025-12-04T08:53:58.7748264Z * [new branch] gh/tugsbayasgalan/55/orig -> origin/gh/tugsbayasgalan/55/orig 2025-12-04T08:53:58.7748346Z * [new branch] gh/tugsbayasgalan/59/base -> origin/gh/tugsbayasgalan/59/base 2025-12-04T08:53:58.7748430Z * [new branch] gh/tugsbayasgalan/59/head -> origin/gh/tugsbayasgalan/59/head 2025-12-04T08:53:58.7748513Z * [new branch] gh/tugsbayasgalan/59/orig -> origin/gh/tugsbayasgalan/59/orig 2025-12-04T08:53:58.7748595Z * [new branch] gh/tugsbayasgalan/6/base -> origin/gh/tugsbayasgalan/6/base 2025-12-04T08:53:58.7748721Z * [new branch] gh/tugsbayasgalan/6/head -> origin/gh/tugsbayasgalan/6/head 2025-12-04T08:53:58.7748803Z * [new branch] gh/tugsbayasgalan/6/orig -> origin/gh/tugsbayasgalan/6/orig 2025-12-04T08:53:58.7748889Z * [new branch] gh/tugsbayasgalan/60/base -> origin/gh/tugsbayasgalan/60/base 2025-12-04T08:53:58.7748972Z * [new branch] gh/tugsbayasgalan/60/head -> origin/gh/tugsbayasgalan/60/head 2025-12-04T08:53:58.7749056Z * [new branch] gh/tugsbayasgalan/60/orig -> origin/gh/tugsbayasgalan/60/orig 2025-12-04T08:53:58.7749139Z * [new branch] gh/tugsbayasgalan/61/base -> origin/gh/tugsbayasgalan/61/base 2025-12-04T08:53:58.7749221Z * [new branch] gh/tugsbayasgalan/61/head -> origin/gh/tugsbayasgalan/61/head 2025-12-04T08:53:58.7749304Z * [new branch] gh/tugsbayasgalan/61/orig -> origin/gh/tugsbayasgalan/61/orig 2025-12-04T08:53:58.7749387Z * [new branch] gh/tugsbayasgalan/63/base -> origin/gh/tugsbayasgalan/63/base 2025-12-04T08:53:58.7749471Z * [new branch] gh/tugsbayasgalan/63/head -> origin/gh/tugsbayasgalan/63/head 2025-12-04T08:53:58.7749555Z * [new branch] gh/tugsbayasgalan/63/orig -> origin/gh/tugsbayasgalan/63/orig 2025-12-04T08:53:58.7749636Z * [new branch] gh/tugsbayasgalan/67/base -> origin/gh/tugsbayasgalan/67/base 2025-12-04T08:53:58.7749742Z * [new branch] gh/tugsbayasgalan/67/head -> origin/gh/tugsbayasgalan/67/head 2025-12-04T08:53:58.7749830Z * [new branch] gh/tugsbayasgalan/67/orig -> origin/gh/tugsbayasgalan/67/orig 2025-12-04T08:53:58.7749913Z * [new branch] gh/tugsbayasgalan/68/base -> origin/gh/tugsbayasgalan/68/base 2025-12-04T08:53:58.7749995Z * [new branch] gh/tugsbayasgalan/68/head -> origin/gh/tugsbayasgalan/68/head 2025-12-04T08:53:58.7750078Z * [new branch] gh/tugsbayasgalan/68/orig -> origin/gh/tugsbayasgalan/68/orig 2025-12-04T08:53:58.7750161Z * [new branch] gh/tugsbayasgalan/7/base -> origin/gh/tugsbayasgalan/7/base 2025-12-04T08:53:58.7750242Z * [new branch] gh/tugsbayasgalan/7/head -> origin/gh/tugsbayasgalan/7/head 2025-12-04T08:53:58.7750323Z * [new branch] gh/tugsbayasgalan/7/orig -> origin/gh/tugsbayasgalan/7/orig 2025-12-04T08:53:58.7750407Z * [new branch] gh/tugsbayasgalan/70/base -> origin/gh/tugsbayasgalan/70/base 2025-12-04T08:53:58.7750489Z * [new branch] gh/tugsbayasgalan/70/head -> origin/gh/tugsbayasgalan/70/head 2025-12-04T08:53:58.7750575Z * [new branch] gh/tugsbayasgalan/70/orig -> origin/gh/tugsbayasgalan/70/orig 2025-12-04T08:53:58.7750658Z * [new branch] gh/tugsbayasgalan/71/base -> origin/gh/tugsbayasgalan/71/base 2025-12-04T08:53:58.7750741Z * [new branch] gh/tugsbayasgalan/71/head -> origin/gh/tugsbayasgalan/71/head 2025-12-04T08:53:58.7750823Z * [new branch] gh/tugsbayasgalan/71/orig -> origin/gh/tugsbayasgalan/71/orig 2025-12-04T08:53:58.7750908Z * [new branch] gh/tugsbayasgalan/72/base -> origin/gh/tugsbayasgalan/72/base 2025-12-04T08:53:58.7750991Z * [new branch] gh/tugsbayasgalan/72/head -> origin/gh/tugsbayasgalan/72/head 2025-12-04T08:53:58.7751074Z * [new branch] gh/tugsbayasgalan/72/orig -> origin/gh/tugsbayasgalan/72/orig 2025-12-04T08:53:58.7751157Z * [new branch] gh/tugsbayasgalan/73/base -> origin/gh/tugsbayasgalan/73/base 2025-12-04T08:53:58.7751243Z * [new branch] gh/tugsbayasgalan/73/head -> origin/gh/tugsbayasgalan/73/head 2025-12-04T08:53:58.7751324Z * [new branch] gh/tugsbayasgalan/73/orig -> origin/gh/tugsbayasgalan/73/orig 2025-12-04T08:53:58.7751407Z * [new branch] gh/tugsbayasgalan/74/base -> origin/gh/tugsbayasgalan/74/base 2025-12-04T08:53:58.7751491Z * [new branch] gh/tugsbayasgalan/74/head -> origin/gh/tugsbayasgalan/74/head 2025-12-04T08:53:58.7751602Z * [new branch] gh/tugsbayasgalan/74/orig -> origin/gh/tugsbayasgalan/74/orig 2025-12-04T08:53:58.7751684Z * [new branch] gh/tugsbayasgalan/75/base -> origin/gh/tugsbayasgalan/75/base 2025-12-04T08:53:58.7751767Z * [new branch] gh/tugsbayasgalan/75/head -> origin/gh/tugsbayasgalan/75/head 2025-12-04T08:53:58.7751852Z * [new branch] gh/tugsbayasgalan/75/orig -> origin/gh/tugsbayasgalan/75/orig 2025-12-04T08:53:58.7751934Z * [new branch] gh/tugsbayasgalan/76/base -> origin/gh/tugsbayasgalan/76/base 2025-12-04T08:53:58.7752017Z * [new branch] gh/tugsbayasgalan/76/head -> origin/gh/tugsbayasgalan/76/head 2025-12-04T08:53:58.7752100Z * [new branch] gh/tugsbayasgalan/76/orig -> origin/gh/tugsbayasgalan/76/orig 2025-12-04T08:53:58.7752182Z * [new branch] gh/tugsbayasgalan/77/base -> origin/gh/tugsbayasgalan/77/base 2025-12-04T08:53:58.7752267Z * [new branch] gh/tugsbayasgalan/77/head -> origin/gh/tugsbayasgalan/77/head 2025-12-04T08:53:58.7752350Z * [new branch] gh/tugsbayasgalan/77/orig -> origin/gh/tugsbayasgalan/77/orig 2025-12-04T08:53:58.7752432Z * [new branch] gh/tugsbayasgalan/78/base -> origin/gh/tugsbayasgalan/78/base 2025-12-04T08:53:58.7752515Z * [new branch] gh/tugsbayasgalan/78/head -> origin/gh/tugsbayasgalan/78/head 2025-12-04T08:53:58.7752619Z * [new branch] gh/tugsbayasgalan/78/orig -> origin/gh/tugsbayasgalan/78/orig 2025-12-04T08:53:58.7752702Z * [new branch] gh/tugsbayasgalan/79/base -> origin/gh/tugsbayasgalan/79/base 2025-12-04T08:53:58.7752784Z * [new branch] gh/tugsbayasgalan/79/head -> origin/gh/tugsbayasgalan/79/head 2025-12-04T08:53:58.7752866Z * [new branch] gh/tugsbayasgalan/79/orig -> origin/gh/tugsbayasgalan/79/orig 2025-12-04T08:53:58.7752949Z * [new branch] gh/tugsbayasgalan/8/base -> origin/gh/tugsbayasgalan/8/base 2025-12-04T08:53:58.7753031Z * [new branch] gh/tugsbayasgalan/8/head -> origin/gh/tugsbayasgalan/8/head 2025-12-04T08:53:58.7753111Z * [new branch] gh/tugsbayasgalan/8/orig -> origin/gh/tugsbayasgalan/8/orig 2025-12-04T08:53:58.7753195Z * [new branch] gh/tugsbayasgalan/80/base -> origin/gh/tugsbayasgalan/80/base 2025-12-04T08:53:58.7753280Z * [new branch] gh/tugsbayasgalan/80/head -> origin/gh/tugsbayasgalan/80/head 2025-12-04T08:53:58.7753363Z * [new branch] gh/tugsbayasgalan/80/orig -> origin/gh/tugsbayasgalan/80/orig 2025-12-04T08:53:58.7753446Z * [new branch] gh/tugsbayasgalan/81/base -> origin/gh/tugsbayasgalan/81/base 2025-12-04T08:53:58.7753528Z * [new branch] gh/tugsbayasgalan/81/head -> origin/gh/tugsbayasgalan/81/head 2025-12-04T08:53:58.7753609Z * [new branch] gh/tugsbayasgalan/81/orig -> origin/gh/tugsbayasgalan/81/orig 2025-12-04T08:53:58.7753694Z * [new branch] gh/tugsbayasgalan/82/base -> origin/gh/tugsbayasgalan/82/base 2025-12-04T08:53:58.7753776Z * [new branch] gh/tugsbayasgalan/82/head -> origin/gh/tugsbayasgalan/82/head 2025-12-04T08:53:58.7753857Z * [new branch] gh/tugsbayasgalan/82/orig -> origin/gh/tugsbayasgalan/82/orig 2025-12-04T08:53:58.7753942Z * [new branch] gh/tugsbayasgalan/83/base -> origin/gh/tugsbayasgalan/83/base 2025-12-04T08:53:58.7754024Z * [new branch] gh/tugsbayasgalan/83/head -> origin/gh/tugsbayasgalan/83/head 2025-12-04T08:53:58.7754108Z * [new branch] gh/tugsbayasgalan/83/orig -> origin/gh/tugsbayasgalan/83/orig 2025-12-04T08:53:58.7754189Z * [new branch] gh/tugsbayasgalan/84/base -> origin/gh/tugsbayasgalan/84/base 2025-12-04T08:53:58.7754271Z * [new branch] gh/tugsbayasgalan/84/head -> origin/gh/tugsbayasgalan/84/head 2025-12-04T08:53:58.7754354Z * [new branch] gh/tugsbayasgalan/84/orig -> origin/gh/tugsbayasgalan/84/orig 2025-12-04T08:53:58.7754461Z * [new branch] gh/tugsbayasgalan/85/base -> origin/gh/tugsbayasgalan/85/base 2025-12-04T08:53:58.7754542Z * [new branch] gh/tugsbayasgalan/85/head -> origin/gh/tugsbayasgalan/85/head 2025-12-04T08:53:58.7754625Z * [new branch] gh/tugsbayasgalan/85/orig -> origin/gh/tugsbayasgalan/85/orig 2025-12-04T08:53:58.7754709Z * [new branch] gh/tugsbayasgalan/86/base -> origin/gh/tugsbayasgalan/86/base 2025-12-04T08:53:58.7754792Z * [new branch] gh/tugsbayasgalan/86/head -> origin/gh/tugsbayasgalan/86/head 2025-12-04T08:53:58.7754875Z * [new branch] gh/tugsbayasgalan/86/orig -> origin/gh/tugsbayasgalan/86/orig 2025-12-04T08:53:58.7754958Z * [new branch] gh/tugsbayasgalan/87/base -> origin/gh/tugsbayasgalan/87/base 2025-12-04T08:53:58.7755040Z * [new branch] gh/tugsbayasgalan/87/head -> origin/gh/tugsbayasgalan/87/head 2025-12-04T08:53:58.7755125Z * [new branch] gh/tugsbayasgalan/87/orig -> origin/gh/tugsbayasgalan/87/orig 2025-12-04T08:53:58.7755208Z * [new branch] gh/tugsbayasgalan/88/base -> origin/gh/tugsbayasgalan/88/base 2025-12-04T08:53:58.7755290Z * [new branch] gh/tugsbayasgalan/88/head -> origin/gh/tugsbayasgalan/88/head 2025-12-04T08:53:58.7755407Z * [new branch] gh/tugsbayasgalan/88/orig -> origin/gh/tugsbayasgalan/88/orig 2025-12-04T08:53:58.7755490Z * [new branch] gh/tugsbayasgalan/89/base -> origin/gh/tugsbayasgalan/89/base 2025-12-04T08:53:58.7755572Z * [new branch] gh/tugsbayasgalan/89/head -> origin/gh/tugsbayasgalan/89/head 2025-12-04T08:53:58.7755656Z * [new branch] gh/tugsbayasgalan/89/orig -> origin/gh/tugsbayasgalan/89/orig 2025-12-04T08:53:58.7755737Z * [new branch] gh/tugsbayasgalan/9/base -> origin/gh/tugsbayasgalan/9/base 2025-12-04T08:53:58.7755820Z * [new branch] gh/tugsbayasgalan/9/head -> origin/gh/tugsbayasgalan/9/head 2025-12-04T08:53:58.7755900Z * [new branch] gh/tugsbayasgalan/9/orig -> origin/gh/tugsbayasgalan/9/orig 2025-12-04T08:53:58.7755982Z * [new branch] gh/tugsbayasgalan/90/base -> origin/gh/tugsbayasgalan/90/base 2025-12-04T08:53:58.7756066Z * [new branch] gh/tugsbayasgalan/90/head -> origin/gh/tugsbayasgalan/90/head 2025-12-04T08:53:58.7756150Z * [new branch] gh/tugsbayasgalan/90/orig -> origin/gh/tugsbayasgalan/90/orig 2025-12-04T08:53:58.7756231Z * [new branch] gh/tugsbayasgalan/91/base -> origin/gh/tugsbayasgalan/91/base 2025-12-04T08:53:58.7756315Z * [new branch] gh/tugsbayasgalan/91/head -> origin/gh/tugsbayasgalan/91/head 2025-12-04T08:53:58.7756397Z * [new branch] gh/tugsbayasgalan/91/orig -> origin/gh/tugsbayasgalan/91/orig 2025-12-04T08:53:58.7756480Z * [new branch] gh/tugsbayasgalan/92/base -> origin/gh/tugsbayasgalan/92/base 2025-12-04T08:53:58.7756565Z * [new branch] gh/tugsbayasgalan/92/head -> origin/gh/tugsbayasgalan/92/head 2025-12-04T08:53:58.7756647Z * [new branch] gh/tugsbayasgalan/92/orig -> origin/gh/tugsbayasgalan/92/orig 2025-12-04T08:53:58.7756729Z * [new branch] gh/tugsbayasgalan/93/base -> origin/gh/tugsbayasgalan/93/base 2025-12-04T08:53:58.7756866Z * [new branch] gh/tugsbayasgalan/93/head -> origin/gh/tugsbayasgalan/93/head 2025-12-04T08:53:58.7756950Z * [new branch] gh/tugsbayasgalan/93/orig -> origin/gh/tugsbayasgalan/93/orig 2025-12-04T08:53:58.7757017Z * [new branch] gh/v0i0/14/base -> origin/gh/v0i0/14/base 2025-12-04T08:53:58.7757084Z * [new branch] gh/v0i0/14/head -> origin/gh/v0i0/14/head 2025-12-04T08:53:58.7757148Z * [new branch] gh/v0i0/14/orig -> origin/gh/v0i0/14/orig 2025-12-04T08:53:58.7757212Z * [new branch] gh/v0i0/15/base -> origin/gh/v0i0/15/base 2025-12-04T08:53:58.7757317Z * [new branch] gh/v0i0/15/head -> origin/gh/v0i0/15/head 2025-12-04T08:53:58.7757381Z * [new branch] gh/v0i0/15/orig -> origin/gh/v0i0/15/orig 2025-12-04T08:53:58.7757444Z * [new branch] gh/v0i0/16/base -> origin/gh/v0i0/16/base 2025-12-04T08:53:58.7757508Z * [new branch] gh/v0i0/16/head -> origin/gh/v0i0/16/head 2025-12-04T08:53:58.7757570Z * [new branch] gh/v0i0/16/orig -> origin/gh/v0i0/16/orig 2025-12-04T08:53:58.7757634Z * [new branch] gh/v0i0/17/base -> origin/gh/v0i0/17/base 2025-12-04T08:53:58.7757696Z * [new branch] gh/v0i0/17/head -> origin/gh/v0i0/17/head 2025-12-04T08:53:58.7757758Z * [new branch] gh/v0i0/17/orig -> origin/gh/v0i0/17/orig 2025-12-04T08:53:58.7757821Z * [new branch] gh/v0i0/18/base -> origin/gh/v0i0/18/base 2025-12-04T08:53:58.7757884Z * [new branch] gh/v0i0/18/head -> origin/gh/v0i0/18/head 2025-12-04T08:53:58.7757947Z * [new branch] gh/v0i0/18/orig -> origin/gh/v0i0/18/orig 2025-12-04T08:53:58.7758010Z * [new branch] gh/v0i0/19/base -> origin/gh/v0i0/19/base 2025-12-04T08:53:58.7758112Z * [new branch] gh/v0i0/19/head -> origin/gh/v0i0/19/head 2025-12-04T08:53:58.7758176Z * [new branch] gh/v0i0/19/orig -> origin/gh/v0i0/19/orig 2025-12-04T08:53:58.7758259Z * [new branch] gh/vishal9-team/1/base -> origin/gh/vishal9-team/1/base 2025-12-04T08:53:58.7758336Z * [new branch] gh/vishal9-team/1/head -> origin/gh/vishal9-team/1/head 2025-12-04T08:53:58.7758411Z * [new branch] gh/vishal9-team/2/base -> origin/gh/vishal9-team/2/base 2025-12-04T08:53:58.7758487Z * [new branch] gh/vishal9-team/2/head -> origin/gh/vishal9-team/2/head 2025-12-04T08:53:58.7758562Z * [new branch] gh/vishal9-team/2/orig -> origin/gh/vishal9-team/2/orig 2025-12-04T08:53:58.7758635Z * [new branch] gh/vishal9-team/3/base -> origin/gh/vishal9-team/3/base 2025-12-04T08:53:58.7758710Z * [new branch] gh/vishal9-team/3/head -> origin/gh/vishal9-team/3/head 2025-12-04T08:53:58.7758786Z * [new branch] gh/vishal9-team/3/orig -> origin/gh/vishal9-team/3/orig 2025-12-04T08:53:58.7758860Z * [new branch] gh/vishal9-team/4/base -> origin/gh/vishal9-team/4/base 2025-12-04T08:53:58.7758935Z * [new branch] gh/vishal9-team/4/head -> origin/gh/vishal9-team/4/head 2025-12-04T08:53:58.7759008Z * [new branch] gh/vishal9-team/4/orig -> origin/gh/vishal9-team/4/orig 2025-12-04T08:53:58.7759074Z * [new branch] gh/vkuzo/1/next -> origin/gh/vkuzo/1/next 2025-12-04T08:53:58.7759140Z * [new branch] gh/vkuzo/2/next -> origin/gh/vkuzo/2/next 2025-12-04T08:53:58.7759207Z * [new branch] gh/vkuzo/3/next -> origin/gh/vkuzo/3/next 2025-12-04T08:53:58.7759281Z * [new branch] gh/wconstab/424/base -> origin/gh/wconstab/424/base 2025-12-04T08:53:58.7759354Z * [new branch] gh/wconstab/424/head -> origin/gh/wconstab/424/head 2025-12-04T08:53:58.7759427Z * [new branch] gh/wconstab/424/orig -> origin/gh/wconstab/424/orig 2025-12-04T08:53:58.7759498Z * [new branch] gh/wconstab/435/base -> origin/gh/wconstab/435/base 2025-12-04T08:53:58.7759569Z * [new branch] gh/wconstab/435/head -> origin/gh/wconstab/435/head 2025-12-04T08:53:58.7759639Z * [new branch] gh/wconstab/435/orig -> origin/gh/wconstab/435/orig 2025-12-04T08:53:58.7759710Z * [new branch] gh/wconstab/444/base -> origin/gh/wconstab/444/base 2025-12-04T08:53:58.7759781Z * [new branch] gh/wconstab/444/head -> origin/gh/wconstab/444/head 2025-12-04T08:53:58.7759875Z * [new branch] gh/wconstab/444/orig -> origin/gh/wconstab/444/orig 2025-12-04T08:53:58.7759947Z * [new branch] gh/wconstab/447/base -> origin/gh/wconstab/447/base 2025-12-04T08:53:58.7760018Z * [new branch] gh/wconstab/447/head -> origin/gh/wconstab/447/head 2025-12-04T08:53:58.7760090Z * [new branch] gh/wconstab/447/orig -> origin/gh/wconstab/447/orig 2025-12-04T08:53:58.7760162Z * [new branch] gh/wconstab/448/base -> origin/gh/wconstab/448/base 2025-12-04T08:53:58.7760233Z * [new branch] gh/wconstab/448/head -> origin/gh/wconstab/448/head 2025-12-04T08:53:58.7760304Z * [new branch] gh/wconstab/448/orig -> origin/gh/wconstab/448/orig 2025-12-04T08:53:58.7760376Z * [new branch] gh/wconstab/449/base -> origin/gh/wconstab/449/base 2025-12-04T08:53:58.7760446Z * [new branch] gh/wconstab/449/head -> origin/gh/wconstab/449/head 2025-12-04T08:53:58.7760518Z * [new branch] gh/wconstab/449/orig -> origin/gh/wconstab/449/orig 2025-12-04T08:53:58.7760590Z * [new branch] gh/wconstab/450/base -> origin/gh/wconstab/450/base 2025-12-04T08:53:58.7760661Z * [new branch] gh/wconstab/450/head -> origin/gh/wconstab/450/head 2025-12-04T08:53:58.7760756Z * [new branch] gh/wconstab/450/orig -> origin/gh/wconstab/450/orig 2025-12-04T08:53:58.7760827Z * [new branch] gh/wconstab/451/base -> origin/gh/wconstab/451/base 2025-12-04T08:53:58.7760897Z * [new branch] gh/wconstab/451/head -> origin/gh/wconstab/451/head 2025-12-04T08:53:58.7760969Z * [new branch] gh/wconstab/451/orig -> origin/gh/wconstab/451/orig 2025-12-04T08:53:58.7761039Z * [new branch] gh/wconstab/452/base -> origin/gh/wconstab/452/base 2025-12-04T08:53:58.7761110Z * [new branch] gh/wconstab/452/head -> origin/gh/wconstab/452/head 2025-12-04T08:53:58.7761184Z * [new branch] gh/wconstab/452/orig -> origin/gh/wconstab/452/orig 2025-12-04T08:53:58.7761254Z * [new branch] gh/wconstab/453/base -> origin/gh/wconstab/453/base 2025-12-04T08:53:58.7761325Z * [new branch] gh/wconstab/453/head -> origin/gh/wconstab/453/head 2025-12-04T08:53:58.7761399Z * [new branch] gh/wconstab/453/orig -> origin/gh/wconstab/453/orig 2025-12-04T08:53:58.7761469Z * [new branch] gh/wconstab/454/base -> origin/gh/wconstab/454/base 2025-12-04T08:53:58.7761540Z * [new branch] gh/wconstab/454/head -> origin/gh/wconstab/454/head 2025-12-04T08:53:58.7761611Z * [new branch] gh/wconstab/454/orig -> origin/gh/wconstab/454/orig 2025-12-04T08:53:58.7761681Z * [new branch] gh/wconstab/455/base -> origin/gh/wconstab/455/base 2025-12-04T08:53:58.7761751Z * [new branch] gh/wconstab/455/head -> origin/gh/wconstab/455/head 2025-12-04T08:53:58.7761826Z * [new branch] gh/wconstab/455/orig -> origin/gh/wconstab/455/orig 2025-12-04T08:53:58.7761896Z * [new branch] gh/wconstab/456/base -> origin/gh/wconstab/456/base 2025-12-04T08:53:58.7761966Z * [new branch] gh/wconstab/456/head -> origin/gh/wconstab/456/head 2025-12-04T08:53:58.7762041Z * [new branch] gh/wconstab/456/orig -> origin/gh/wconstab/456/orig 2025-12-04T08:53:58.7762112Z * [new branch] gh/wconstab/457/base -> origin/gh/wconstab/457/base 2025-12-04T08:53:58.7762184Z * [new branch] gh/wconstab/457/head -> origin/gh/wconstab/457/head 2025-12-04T08:53:58.7762253Z * [new branch] gh/wconstab/457/orig -> origin/gh/wconstab/457/orig 2025-12-04T08:53:58.7762324Z * [new branch] gh/wconstab/458/base -> origin/gh/wconstab/458/base 2025-12-04T08:53:58.7762429Z * [new branch] gh/wconstab/458/head -> origin/gh/wconstab/458/head 2025-12-04T08:53:58.7762499Z * [new branch] gh/wconstab/458/orig -> origin/gh/wconstab/458/orig 2025-12-04T08:53:58.7762569Z * [new branch] gh/wconstab/459/base -> origin/gh/wconstab/459/base 2025-12-04T08:53:58.7762640Z * [new branch] gh/wconstab/459/head -> origin/gh/wconstab/459/head 2025-12-04T08:53:58.7762712Z * [new branch] gh/wconstab/459/orig -> origin/gh/wconstab/459/orig 2025-12-04T08:53:58.7762783Z * [new branch] gh/wconstab/460/base -> origin/gh/wconstab/460/base 2025-12-04T08:53:58.7762856Z * [new branch] gh/wconstab/460/head -> origin/gh/wconstab/460/head 2025-12-04T08:53:58.7762926Z * [new branch] gh/wconstab/460/orig -> origin/gh/wconstab/460/orig 2025-12-04T08:53:58.7762998Z * [new branch] gh/wconstab/461/base -> origin/gh/wconstab/461/base 2025-12-04T08:53:58.7763071Z * [new branch] gh/wconstab/461/head -> origin/gh/wconstab/461/head 2025-12-04T08:53:58.7763144Z * [new branch] gh/wconstab/461/orig -> origin/gh/wconstab/461/orig 2025-12-04T08:53:58.7763214Z * [new branch] gh/wconstab/462/base -> origin/gh/wconstab/462/base 2025-12-04T08:53:58.7763286Z * [new branch] gh/wconstab/462/head -> origin/gh/wconstab/462/head 2025-12-04T08:53:58.7763385Z * [new branch] gh/wconstab/462/orig -> origin/gh/wconstab/462/orig 2025-12-04T08:53:58.7763457Z * [new branch] gh/wconstab/463/base -> origin/gh/wconstab/463/base 2025-12-04T08:53:58.7763532Z * [new branch] gh/wconstab/463/head -> origin/gh/wconstab/463/head 2025-12-04T08:53:58.7763603Z * [new branch] gh/wconstab/463/orig -> origin/gh/wconstab/463/orig 2025-12-04T08:53:58.7763679Z * [new branch] gh/wconstab/464/base -> origin/gh/wconstab/464/base 2025-12-04T08:53:58.7763753Z * [new branch] gh/wconstab/464/head -> origin/gh/wconstab/464/head 2025-12-04T08:53:58.7763823Z * [new branch] gh/wconstab/464/orig -> origin/gh/wconstab/464/orig 2025-12-04T08:53:58.7763896Z * [new branch] gh/wconstab/465/base -> origin/gh/wconstab/465/base 2025-12-04T08:53:58.7763966Z * [new branch] gh/wconstab/465/head -> origin/gh/wconstab/465/head 2025-12-04T08:53:58.7764037Z * [new branch] gh/wconstab/465/orig -> origin/gh/wconstab/465/orig 2025-12-04T08:53:58.7764111Z * [new branch] gh/wconstab/466/base -> origin/gh/wconstab/466/base 2025-12-04T08:53:58.7764182Z * [new branch] gh/wconstab/466/head -> origin/gh/wconstab/466/head 2025-12-04T08:53:58.7764252Z * [new branch] gh/wconstab/466/orig -> origin/gh/wconstab/466/orig 2025-12-04T08:53:58.7764327Z * [new branch] gh/wconstab/467/base -> origin/gh/wconstab/467/base 2025-12-04T08:53:58.7764398Z * [new branch] gh/wconstab/467/head -> origin/gh/wconstab/467/head 2025-12-04T08:53:58.7764469Z * [new branch] gh/wconstab/467/orig -> origin/gh/wconstab/467/orig 2025-12-04T08:53:58.7764542Z * [new branch] gh/wconstab/468/base -> origin/gh/wconstab/468/base 2025-12-04T08:53:58.7764613Z * [new branch] gh/wconstab/468/head -> origin/gh/wconstab/468/head 2025-12-04T08:53:58.7764687Z * [new branch] gh/wconstab/468/orig -> origin/gh/wconstab/468/orig 2025-12-04T08:53:58.7764764Z * [new branch] gh/weifengpy/39/base -> origin/gh/weifengpy/39/base 2025-12-04T08:53:58.7764837Z * [new branch] gh/weifengpy/39/head -> origin/gh/weifengpy/39/head 2025-12-04T08:53:58.7764912Z * [new branch] gh/weifengpy/39/orig -> origin/gh/weifengpy/39/orig 2025-12-04T08:53:58.7764987Z * [new branch] gh/weifengpy/40/base -> origin/gh/weifengpy/40/base 2025-12-04T08:53:58.7765083Z * [new branch] gh/weifengpy/40/head -> origin/gh/weifengpy/40/head 2025-12-04T08:53:58.7765155Z * [new branch] gh/weifengpy/40/orig -> origin/gh/weifengpy/40/orig 2025-12-04T08:53:58.7765230Z * [new branch] gh/weifengpy/41/base -> origin/gh/weifengpy/41/base 2025-12-04T08:53:58.7765304Z * [new branch] gh/weifengpy/41/head -> origin/gh/weifengpy/41/head 2025-12-04T08:53:58.7765379Z * [new branch] gh/weifengpy/41/orig -> origin/gh/weifengpy/41/orig 2025-12-04T08:53:58.7765463Z * [new branch] gh/williamwen42/250/base -> origin/gh/williamwen42/250/base 2025-12-04T08:53:58.7765545Z * [new branch] gh/williamwen42/250/head -> origin/gh/williamwen42/250/head 2025-12-04T08:53:58.7765627Z * [new branch] gh/williamwen42/250/orig -> origin/gh/williamwen42/250/orig 2025-12-04T08:53:58.7765707Z * [new branch] gh/williamwen42/279/base -> origin/gh/williamwen42/279/base 2025-12-04T08:53:58.7765787Z * [new branch] gh/williamwen42/279/head -> origin/gh/williamwen42/279/head 2025-12-04T08:53:58.7765866Z * [new branch] gh/williamwen42/279/orig -> origin/gh/williamwen42/279/orig 2025-12-04T08:53:58.7765943Z * [new branch] gh/williamwen42/282/base -> origin/gh/williamwen42/282/base 2025-12-04T08:53:58.7766051Z * [new branch] gh/williamwen42/282/head -> origin/gh/williamwen42/282/head 2025-12-04T08:53:58.7766132Z * [new branch] gh/williamwen42/282/orig -> origin/gh/williamwen42/282/orig 2025-12-04T08:53:58.7766212Z * [new branch] gh/williamwen42/287/base -> origin/gh/williamwen42/287/base 2025-12-04T08:53:58.7766290Z * [new branch] gh/williamwen42/287/head -> origin/gh/williamwen42/287/head 2025-12-04T08:53:58.7766372Z * [new branch] gh/williamwen42/287/orig -> origin/gh/williamwen42/287/orig 2025-12-04T08:53:58.7766450Z * [new branch] gh/williamwen42/288/base -> origin/gh/williamwen42/288/base 2025-12-04T08:53:58.7766529Z * [new branch] gh/williamwen42/288/head -> origin/gh/williamwen42/288/head 2025-12-04T08:53:58.7766609Z * [new branch] gh/williamwen42/288/orig -> origin/gh/williamwen42/288/orig 2025-12-04T08:53:58.7766687Z * [new branch] gh/williamwen42/296/base -> origin/gh/williamwen42/296/base 2025-12-04T08:53:58.7766813Z * [new branch] gh/williamwen42/296/head -> origin/gh/williamwen42/296/head 2025-12-04T08:53:58.7766894Z * [new branch] gh/williamwen42/296/orig -> origin/gh/williamwen42/296/orig 2025-12-04T08:53:58.7766973Z * [new branch] gh/williamwen42/297/base -> origin/gh/williamwen42/297/base 2025-12-04T08:53:58.7767055Z * [new branch] gh/williamwen42/297/head -> origin/gh/williamwen42/297/head 2025-12-04T08:53:58.7767134Z * [new branch] gh/williamwen42/297/orig -> origin/gh/williamwen42/297/orig 2025-12-04T08:53:58.7767214Z * [new branch] gh/williamwen42/306/base -> origin/gh/williamwen42/306/base 2025-12-04T08:53:58.7767299Z * [new branch] gh/williamwen42/306/head -> origin/gh/williamwen42/306/head 2025-12-04T08:53:58.7767378Z * [new branch] gh/williamwen42/306/orig -> origin/gh/williamwen42/306/orig 2025-12-04T08:53:58.7767458Z * [new branch] gh/williamwen42/309/base -> origin/gh/williamwen42/309/base 2025-12-04T08:53:58.7767537Z * [new branch] gh/williamwen42/309/head -> origin/gh/williamwen42/309/head 2025-12-04T08:53:58.7767617Z * [new branch] gh/williamwen42/309/orig -> origin/gh/williamwen42/309/orig 2025-12-04T08:53:58.7767697Z * [new branch] gh/williamwen42/310/base -> origin/gh/williamwen42/310/base 2025-12-04T08:53:58.7767778Z * [new branch] gh/williamwen42/310/head -> origin/gh/williamwen42/310/head 2025-12-04T08:53:58.7767899Z * [new branch] gh/williamwen42/310/orig -> origin/gh/williamwen42/310/orig 2025-12-04T08:53:58.7767976Z * [new branch] gh/williamwen42/311/base -> origin/gh/williamwen42/311/base 2025-12-04T08:53:58.7768057Z * [new branch] gh/williamwen42/311/head -> origin/gh/williamwen42/311/head 2025-12-04T08:53:58.7768137Z * [new branch] gh/williamwen42/311/orig -> origin/gh/williamwen42/311/orig 2025-12-04T08:53:58.7768216Z * [new branch] gh/williamwen42/319/base -> origin/gh/williamwen42/319/base 2025-12-04T08:53:58.7768297Z * [new branch] gh/williamwen42/319/head -> origin/gh/williamwen42/319/head 2025-12-04T08:53:58.7768377Z * [new branch] gh/williamwen42/319/orig -> origin/gh/williamwen42/319/orig 2025-12-04T08:53:58.7768457Z * [new branch] gh/williamwen42/325/base -> origin/gh/williamwen42/325/base 2025-12-04T08:53:58.7768537Z * [new branch] gh/williamwen42/325/head -> origin/gh/williamwen42/325/head 2025-12-04T08:53:58.7768618Z * [new branch] gh/williamwen42/325/orig -> origin/gh/williamwen42/325/orig 2025-12-04T08:53:58.7768699Z * [new branch] gh/williamwen42/326/base -> origin/gh/williamwen42/326/base 2025-12-04T08:53:58.7768778Z * [new branch] gh/williamwen42/326/head -> origin/gh/williamwen42/326/head 2025-12-04T08:53:58.7768894Z * [new branch] gh/williamwen42/326/orig -> origin/gh/williamwen42/326/orig 2025-12-04T08:53:58.7768976Z * [new branch] gh/williamwen42/327/base -> origin/gh/williamwen42/327/base 2025-12-04T08:53:58.7769055Z * [new branch] gh/williamwen42/327/head -> origin/gh/williamwen42/327/head 2025-12-04T08:53:58.7769133Z * [new branch] gh/williamwen42/327/orig -> origin/gh/williamwen42/327/orig 2025-12-04T08:53:58.7769215Z * [new branch] gh/williamwen42/328/base -> origin/gh/williamwen42/328/base 2025-12-04T08:53:58.7769296Z * [new branch] gh/williamwen42/328/head -> origin/gh/williamwen42/328/head 2025-12-04T08:53:58.7769374Z * [new branch] gh/williamwen42/328/orig -> origin/gh/williamwen42/328/orig 2025-12-04T08:53:58.7769454Z * [new branch] gh/williamwen42/329/base -> origin/gh/williamwen42/329/base 2025-12-04T08:53:58.7769532Z * [new branch] gh/williamwen42/329/head -> origin/gh/williamwen42/329/head 2025-12-04T08:53:58.7769612Z * [new branch] gh/williamwen42/329/orig -> origin/gh/williamwen42/329/orig 2025-12-04T08:53:58.7769690Z * [new branch] gh/williamwen42/330/base -> origin/gh/williamwen42/330/base 2025-12-04T08:53:58.7769768Z * [new branch] gh/williamwen42/330/head -> origin/gh/williamwen42/330/head 2025-12-04T08:53:58.7769847Z * [new branch] gh/williamwen42/330/orig -> origin/gh/williamwen42/330/orig 2025-12-04T08:53:58.7769927Z * [new branch] gh/williamwen42/331/base -> origin/gh/williamwen42/331/base 2025-12-04T08:53:58.7770008Z * [new branch] gh/williamwen42/331/head -> origin/gh/williamwen42/331/head 2025-12-04T08:53:58.7770089Z * [new branch] gh/williamwen42/331/orig -> origin/gh/williamwen42/331/orig 2025-12-04T08:53:58.7770168Z * [new branch] gh/williamwen42/332/base -> origin/gh/williamwen42/332/base 2025-12-04T08:53:58.7770249Z * [new branch] gh/williamwen42/332/head -> origin/gh/williamwen42/332/head 2025-12-04T08:53:58.7770331Z * [new branch] gh/williamwen42/332/orig -> origin/gh/williamwen42/332/orig 2025-12-04T08:53:58.7770408Z * [new branch] gh/williamwen42/333/base -> origin/gh/williamwen42/333/base 2025-12-04T08:53:58.7770487Z * [new branch] gh/williamwen42/333/head -> origin/gh/williamwen42/333/head 2025-12-04T08:53:58.7770569Z * [new branch] gh/williamwen42/333/orig -> origin/gh/williamwen42/333/orig 2025-12-04T08:53:58.7770648Z * [new branch] gh/williamwen42/334/base -> origin/gh/williamwen42/334/base 2025-12-04T08:53:58.7770756Z * [new branch] gh/williamwen42/334/head -> origin/gh/williamwen42/334/head 2025-12-04T08:53:58.7770837Z * [new branch] gh/williamwen42/334/orig -> origin/gh/williamwen42/334/orig 2025-12-04T08:53:58.7770915Z * [new branch] gh/williamwen42/335/base -> origin/gh/williamwen42/335/base 2025-12-04T08:53:58.7770994Z * [new branch] gh/williamwen42/335/head -> origin/gh/williamwen42/335/head 2025-12-04T08:53:58.7771074Z * [new branch] gh/williamwen42/335/orig -> origin/gh/williamwen42/335/orig 2025-12-04T08:53:58.7771153Z * [new branch] gh/williamwen42/336/base -> origin/gh/williamwen42/336/base 2025-12-04T08:53:58.7771232Z * [new branch] gh/williamwen42/336/head -> origin/gh/williamwen42/336/head 2025-12-04T08:53:58.7771312Z * [new branch] gh/williamwen42/336/orig -> origin/gh/williamwen42/336/orig 2025-12-04T08:53:58.7771393Z * [new branch] gh/williamwen42/337/base -> origin/gh/williamwen42/337/base 2025-12-04T08:53:58.7771474Z * [new branch] gh/williamwen42/337/head -> origin/gh/williamwen42/337/head 2025-12-04T08:53:58.7771552Z * [new branch] gh/williamwen42/337/orig -> origin/gh/williamwen42/337/orig 2025-12-04T08:53:58.7771652Z * [new branch] gh/williamwen42/338/base -> origin/gh/williamwen42/338/base 2025-12-04T08:53:58.7771733Z * [new branch] gh/williamwen42/338/head -> origin/gh/williamwen42/338/head 2025-12-04T08:53:58.7771810Z * [new branch] gh/williamwen42/338/orig -> origin/gh/williamwen42/338/orig 2025-12-04T08:53:58.7771888Z * [new branch] gh/williamwen42/339/base -> origin/gh/williamwen42/339/base 2025-12-04T08:53:58.7771969Z * [new branch] gh/williamwen42/339/head -> origin/gh/williamwen42/339/head 2025-12-04T08:53:58.7772047Z * [new branch] gh/williamwen42/339/orig -> origin/gh/williamwen42/339/orig 2025-12-04T08:53:58.7772127Z * [new branch] gh/williamwen42/340/base -> origin/gh/williamwen42/340/base 2025-12-04T08:53:58.7772206Z * [new branch] gh/williamwen42/340/head -> origin/gh/williamwen42/340/head 2025-12-04T08:53:58.7772283Z * [new branch] gh/williamwen42/340/orig -> origin/gh/williamwen42/340/orig 2025-12-04T08:53:58.7772361Z * [new branch] gh/williamwen42/341/base -> origin/gh/williamwen42/341/base 2025-12-04T08:53:58.7772440Z * [new branch] gh/williamwen42/341/head -> origin/gh/williamwen42/341/head 2025-12-04T08:53:58.7772518Z * [new branch] gh/williamwen42/341/orig -> origin/gh/williamwen42/341/orig 2025-12-04T08:53:58.7772596Z * [new branch] gh/williamwen42/342/base -> origin/gh/williamwen42/342/base 2025-12-04T08:53:58.7772675Z * [new branch] gh/williamwen42/342/head -> origin/gh/williamwen42/342/head 2025-12-04T08:53:58.7772755Z * [new branch] gh/williamwen42/342/orig -> origin/gh/williamwen42/342/orig 2025-12-04T08:53:58.7772835Z * [new branch] gh/williamwen42/343/base -> origin/gh/williamwen42/343/base 2025-12-04T08:53:58.7772912Z * [new branch] gh/williamwen42/343/head -> origin/gh/williamwen42/343/head 2025-12-04T08:53:58.7772992Z * [new branch] gh/williamwen42/343/orig -> origin/gh/williamwen42/343/orig 2025-12-04T08:53:58.7773071Z * [new branch] gh/williamwen42/344/base -> origin/gh/williamwen42/344/base 2025-12-04T08:53:58.7773148Z * [new branch] gh/williamwen42/344/head -> origin/gh/williamwen42/344/head 2025-12-04T08:53:58.7773225Z * [new branch] gh/williamwen42/344/orig -> origin/gh/williamwen42/344/orig 2025-12-04T08:53:58.7773304Z * [new branch] gh/williamwen42/345/base -> origin/gh/williamwen42/345/base 2025-12-04T08:53:58.7773382Z * [new branch] gh/williamwen42/345/head -> origin/gh/williamwen42/345/head 2025-12-04T08:53:58.7773483Z * [new branch] gh/williamwen42/345/orig -> origin/gh/williamwen42/345/orig 2025-12-04T08:53:58.7773562Z * [new branch] gh/williamwen42/346/base -> origin/gh/williamwen42/346/base 2025-12-04T08:53:58.7773639Z * [new branch] gh/williamwen42/346/head -> origin/gh/williamwen42/346/head 2025-12-04T08:53:58.7773719Z * [new branch] gh/williamwen42/346/orig -> origin/gh/williamwen42/346/orig 2025-12-04T08:53:58.7773798Z * [new branch] gh/williamwen42/347/base -> origin/gh/williamwen42/347/base 2025-12-04T08:53:58.7773876Z * [new branch] gh/williamwen42/347/head -> origin/gh/williamwen42/347/head 2025-12-04T08:53:58.7773954Z * [new branch] gh/williamwen42/347/orig -> origin/gh/williamwen42/347/orig 2025-12-04T08:53:58.7774032Z * [new branch] gh/williamwen42/348/base -> origin/gh/williamwen42/348/base 2025-12-04T08:53:58.7774111Z * [new branch] gh/williamwen42/348/head -> origin/gh/williamwen42/348/head 2025-12-04T08:53:58.7774188Z * [new branch] gh/williamwen42/348/orig -> origin/gh/williamwen42/348/orig 2025-12-04T08:53:58.7774267Z * [new branch] gh/williamwen42/349/base -> origin/gh/williamwen42/349/base 2025-12-04T08:53:58.7774371Z * [new branch] gh/williamwen42/349/head -> origin/gh/williamwen42/349/head 2025-12-04T08:53:58.7774451Z * [new branch] gh/williamwen42/349/orig -> origin/gh/williamwen42/349/orig 2025-12-04T08:53:58.7774528Z * [new branch] gh/williamwen42/350/base -> origin/gh/williamwen42/350/base 2025-12-04T08:53:58.7774605Z * [new branch] gh/williamwen42/350/head -> origin/gh/williamwen42/350/head 2025-12-04T08:53:58.7774684Z * [new branch] gh/williamwen42/350/orig -> origin/gh/williamwen42/350/orig 2025-12-04T08:53:58.7774760Z * [new branch] gh/williamwen42/351/base -> origin/gh/williamwen42/351/base 2025-12-04T08:53:58.7774840Z * [new branch] gh/williamwen42/351/head -> origin/gh/williamwen42/351/head 2025-12-04T08:53:58.7774919Z * [new branch] gh/williamwen42/351/orig -> origin/gh/williamwen42/351/orig 2025-12-04T08:53:58.7774997Z * [new branch] gh/williamwen42/352/base -> origin/gh/williamwen42/352/base 2025-12-04T08:53:58.7775076Z * [new branch] gh/williamwen42/352/head -> origin/gh/williamwen42/352/head 2025-12-04T08:53:58.7775154Z * [new branch] gh/williamwen42/352/orig -> origin/gh/williamwen42/352/orig 2025-12-04T08:53:58.7775231Z * [new branch] gh/williamwen42/353/base -> origin/gh/williamwen42/353/base 2025-12-04T08:53:58.7775309Z * [new branch] gh/williamwen42/353/head -> origin/gh/williamwen42/353/head 2025-12-04T08:53:58.7775388Z * [new branch] gh/williamwen42/353/orig -> origin/gh/williamwen42/353/orig 2025-12-04T08:53:58.7775469Z * [new branch] gh/williamwen42/354/base -> origin/gh/williamwen42/354/base 2025-12-04T08:53:58.7775546Z * [new branch] gh/williamwen42/354/head -> origin/gh/williamwen42/354/head 2025-12-04T08:53:58.7775625Z * [new branch] gh/williamwen42/354/orig -> origin/gh/williamwen42/354/orig 2025-12-04T08:53:58.7775704Z * [new branch] gh/williamwen42/355/base -> origin/gh/williamwen42/355/base 2025-12-04T08:53:58.7775783Z * [new branch] gh/williamwen42/355/head -> origin/gh/williamwen42/355/head 2025-12-04T08:53:58.7775861Z * [new branch] gh/williamwen42/355/orig -> origin/gh/williamwen42/355/orig 2025-12-04T08:53:58.7775938Z * [new branch] gh/williamwen42/356/base -> origin/gh/williamwen42/356/base 2025-12-04T08:53:58.7776017Z * [new branch] gh/williamwen42/356/head -> origin/gh/williamwen42/356/head 2025-12-04T08:53:58.7776095Z * [new branch] gh/williamwen42/356/orig -> origin/gh/williamwen42/356/orig 2025-12-04T08:53:58.7776204Z * [new branch] gh/williamwen42/357/base -> origin/gh/williamwen42/357/base 2025-12-04T08:53:58.7776284Z * [new branch] gh/williamwen42/357/head -> origin/gh/williamwen42/357/head 2025-12-04T08:53:58.7776362Z * [new branch] gh/williamwen42/357/orig -> origin/gh/williamwen42/357/orig 2025-12-04T08:53:58.7776440Z * [new branch] gh/williamwen42/358/base -> origin/gh/williamwen42/358/base 2025-12-04T08:53:58.7776519Z * [new branch] gh/williamwen42/358/head -> origin/gh/williamwen42/358/head 2025-12-04T08:53:58.7776597Z * [new branch] gh/williamwen42/358/orig -> origin/gh/williamwen42/358/orig 2025-12-04T08:53:58.7776666Z * [new branch] gh/xmfan/169/base -> origin/gh/xmfan/169/base 2025-12-04T08:53:58.7776737Z * [new branch] gh/xmfan/169/head -> origin/gh/xmfan/169/head 2025-12-04T08:53:58.7776858Z * [new branch] gh/xmfan/170/base -> origin/gh/xmfan/170/base 2025-12-04T08:53:58.7776925Z * [new branch] gh/xmfan/170/head -> origin/gh/xmfan/170/head 2025-12-04T08:53:58.7776994Z * [new branch] gh/xmfan/274/base -> origin/gh/xmfan/274/base 2025-12-04T08:53:58.7777061Z * [new branch] gh/xmfan/274/head -> origin/gh/xmfan/274/head 2025-12-04T08:53:58.7777165Z * [new branch] gh/xmfan/274/orig -> origin/gh/xmfan/274/orig 2025-12-04T08:53:58.7777233Z * [new branch] gh/xmfan/277/base -> origin/gh/xmfan/277/base 2025-12-04T08:53:58.7777300Z * [new branch] gh/xmfan/277/head -> origin/gh/xmfan/277/head 2025-12-04T08:53:58.7777372Z * [new branch] gh/xmfan/277/orig -> origin/gh/xmfan/277/orig 2025-12-04T08:53:58.7777443Z * [new branch] gh/xmfan/301/base -> origin/gh/xmfan/301/base 2025-12-04T08:53:58.7777511Z * [new branch] gh/xmfan/301/head -> origin/gh/xmfan/301/head 2025-12-04T08:53:58.7777580Z * [new branch] gh/xmfan/301/orig -> origin/gh/xmfan/301/orig 2025-12-04T08:53:58.7777648Z * [new branch] gh/xmfan/304/base -> origin/gh/xmfan/304/base 2025-12-04T08:53:58.7777714Z * [new branch] gh/xmfan/304/head -> origin/gh/xmfan/304/head 2025-12-04T08:53:58.7777784Z * [new branch] gh/xmfan/304/orig -> origin/gh/xmfan/304/orig 2025-12-04T08:53:58.7777851Z * [new branch] gh/xmfan/309/base -> origin/gh/xmfan/309/base 2025-12-04T08:53:58.7777917Z * [new branch] gh/xmfan/309/head -> origin/gh/xmfan/309/head 2025-12-04T08:53:58.7777986Z * [new branch] gh/xmfan/309/orig -> origin/gh/xmfan/309/orig 2025-12-04T08:53:58.7778053Z * [new branch] gh/xmfan/310/base -> origin/gh/xmfan/310/base 2025-12-04T08:53:58.7778119Z * [new branch] gh/xmfan/310/head -> origin/gh/xmfan/310/head 2025-12-04T08:53:58.7778189Z * [new branch] gh/xmfan/310/orig -> origin/gh/xmfan/310/orig 2025-12-04T08:53:58.7778256Z * [new branch] gh/xmfan/311/base -> origin/gh/xmfan/311/base 2025-12-04T08:53:58.7778322Z * [new branch] gh/xmfan/311/head -> origin/gh/xmfan/311/head 2025-12-04T08:53:58.7778392Z * [new branch] gh/xmfan/311/orig -> origin/gh/xmfan/311/orig 2025-12-04T08:53:58.7778459Z * [new branch] gh/xmfan/312/base -> origin/gh/xmfan/312/base 2025-12-04T08:53:58.7778525Z * [new branch] gh/xmfan/312/head -> origin/gh/xmfan/312/head 2025-12-04T08:53:58.7778593Z * [new branch] gh/xmfan/312/orig -> origin/gh/xmfan/312/orig 2025-12-04T08:53:58.7778660Z * [new branch] gh/xmfan/313/base -> origin/gh/xmfan/313/base 2025-12-04T08:53:58.7778726Z * [new branch] gh/xmfan/313/head -> origin/gh/xmfan/313/head 2025-12-04T08:53:58.7778830Z * [new branch] gh/xmfan/313/orig -> origin/gh/xmfan/313/orig 2025-12-04T08:53:58.7778910Z * [new branch] gh/xuanzhang816/27/base -> origin/gh/xuanzhang816/27/base 2025-12-04T08:53:58.7778989Z * [new branch] gh/xuanzhang816/27/head -> origin/gh/xuanzhang816/27/head 2025-12-04T08:53:58.7779067Z * [new branch] gh/xuanzhang816/27/orig -> origin/gh/xuanzhang816/27/orig 2025-12-04T08:53:58.7779142Z * [new branch] gh/xuanzhang816/32/base -> origin/gh/xuanzhang816/32/base 2025-12-04T08:53:58.7779220Z * [new branch] gh/xuanzhang816/32/head -> origin/gh/xuanzhang816/32/head 2025-12-04T08:53:58.7779295Z * [new branch] gh/xuanzhang816/32/orig -> origin/gh/xuanzhang816/32/orig 2025-12-04T08:53:58.7779370Z * [new branch] gh/xuanzhang816/33/base -> origin/gh/xuanzhang816/33/base 2025-12-04T08:53:58.7779447Z * [new branch] gh/xuanzhang816/33/head -> origin/gh/xuanzhang816/33/head 2025-12-04T08:53:58.7779524Z * [new branch] gh/xuanzhang816/33/orig -> origin/gh/xuanzhang816/33/orig 2025-12-04T08:53:58.7779599Z * [new branch] gh/xuanzhang816/34/base -> origin/gh/xuanzhang816/34/base 2025-12-04T08:53:58.7779676Z * [new branch] gh/xuanzhang816/34/head -> origin/gh/xuanzhang816/34/head 2025-12-04T08:53:58.7779773Z * [new branch] gh/xuanzhang816/34/orig -> origin/gh/xuanzhang816/34/orig 2025-12-04T08:53:58.7779850Z * [new branch] gh/xuanzhang816/35/base -> origin/gh/xuanzhang816/35/base 2025-12-04T08:53:58.7779925Z * [new branch] gh/xuanzhang816/35/head -> origin/gh/xuanzhang816/35/head 2025-12-04T08:53:58.7780000Z * [new branch] gh/xuanzhang816/35/orig -> origin/gh/xuanzhang816/35/orig 2025-12-04T08:53:58.7780074Z * [new branch] gh/yanbing-j/11/base -> origin/gh/yanbing-j/11/base 2025-12-04T08:53:58.7780150Z * [new branch] gh/yanbing-j/11/head -> origin/gh/yanbing-j/11/head 2025-12-04T08:53:58.7780222Z * [new branch] gh/yanbing-j/11/orig -> origin/gh/yanbing-j/11/orig 2025-12-04T08:53:58.7780292Z * [new branch] gh/yanbing-j/12/base -> origin/gh/yanbing-j/12/base 2025-12-04T08:53:58.7780367Z * [new branch] gh/yanbing-j/12/head -> origin/gh/yanbing-j/12/head 2025-12-04T08:53:58.7780438Z * [new branch] gh/yanbing-j/12/orig -> origin/gh/yanbing-j/12/orig 2025-12-04T08:53:58.7780509Z * [new branch] gh/yanbing-j/13/base -> origin/gh/yanbing-j/13/base 2025-12-04T08:53:58.7780580Z * [new branch] gh/yanbing-j/13/head -> origin/gh/yanbing-j/13/head 2025-12-04T08:53:58.7780650Z * [new branch] gh/yanbing-j/13/orig -> origin/gh/yanbing-j/13/orig 2025-12-04T08:53:58.7780721Z * [new branch] gh/yanbing-j/14/base -> origin/gh/yanbing-j/14/base 2025-12-04T08:53:58.7780793Z * [new branch] gh/yanbing-j/14/head -> origin/gh/yanbing-j/14/head 2025-12-04T08:53:58.7780863Z * [new branch] gh/yanbing-j/14/orig -> origin/gh/yanbing-j/14/orig 2025-12-04T08:53:58.7780934Z * [new branch] gh/yanbing-j/15/base -> origin/gh/yanbing-j/15/base 2025-12-04T08:53:58.7781004Z * [new branch] gh/yanbing-j/15/head -> origin/gh/yanbing-j/15/head 2025-12-04T08:53:58.7781075Z * [new branch] gh/yanbing-j/15/orig -> origin/gh/yanbing-j/15/orig 2025-12-04T08:53:58.7781146Z * [new branch] gh/yanbing-j/18/base -> origin/gh/yanbing-j/18/base 2025-12-04T08:53:58.7781215Z * [new branch] gh/yanbing-j/18/head -> origin/gh/yanbing-j/18/head 2025-12-04T08:53:58.7781286Z * [new branch] gh/yanbing-j/18/orig -> origin/gh/yanbing-j/18/orig 2025-12-04T08:53:58.7781358Z * [new branch] gh/yanbing-j/19/base -> origin/gh/yanbing-j/19/base 2025-12-04T08:53:58.7781474Z * [new branch] gh/yanbing-j/19/head -> origin/gh/yanbing-j/19/head 2025-12-04T08:53:58.7781544Z * [new branch] gh/yanbing-j/19/orig -> origin/gh/yanbing-j/19/orig 2025-12-04T08:53:58.7781616Z * [new branch] gh/yanbing-j/20/base -> origin/gh/yanbing-j/20/base 2025-12-04T08:53:58.7781685Z * [new branch] gh/yanbing-j/20/head -> origin/gh/yanbing-j/20/head 2025-12-04T08:53:58.7781757Z * [new branch] gh/yanbing-j/20/orig -> origin/gh/yanbing-j/20/orig 2025-12-04T08:53:58.7781829Z * [new branch] gh/yanbing-j/21/base -> origin/gh/yanbing-j/21/base 2025-12-04T08:53:58.7781899Z * [new branch] gh/yanbing-j/21/head -> origin/gh/yanbing-j/21/head 2025-12-04T08:53:58.7781970Z * [new branch] gh/yanbing-j/22/base -> origin/gh/yanbing-j/22/base 2025-12-04T08:53:58.7782043Z * [new branch] gh/yanbing-j/22/head -> origin/gh/yanbing-j/22/head 2025-12-04T08:53:58.7782115Z * [new branch] gh/yanbing-j/22/orig -> origin/gh/yanbing-j/22/orig 2025-12-04T08:53:58.7782185Z * [new branch] gh/yanbing-j/23/base -> origin/gh/yanbing-j/23/base 2025-12-04T08:53:58.7782257Z * [new branch] gh/yanbing-j/23/head -> origin/gh/yanbing-j/23/head 2025-12-04T08:53:58.7782354Z * [new branch] gh/yanbing-j/23/orig -> origin/gh/yanbing-j/23/orig 2025-12-04T08:53:58.7782426Z * [new branch] gh/yanbing-j/24/base -> origin/gh/yanbing-j/24/base 2025-12-04T08:53:58.7782497Z * [new branch] gh/yanbing-j/24/head -> origin/gh/yanbing-j/24/head 2025-12-04T08:53:58.7782568Z * [new branch] gh/yanbing-j/24/orig -> origin/gh/yanbing-j/24/orig 2025-12-04T08:53:58.7782640Z * [new branch] gh/yanbing-j/25/base -> origin/gh/yanbing-j/25/base 2025-12-04T08:53:58.7782709Z * [new branch] gh/yanbing-j/25/head -> origin/gh/yanbing-j/25/head 2025-12-04T08:53:58.7782782Z * [new branch] gh/yanbing-j/25/orig -> origin/gh/yanbing-j/25/orig 2025-12-04T08:53:58.7782854Z * [new branch] gh/yanbing-j/26/base -> origin/gh/yanbing-j/26/base 2025-12-04T08:53:58.7782924Z * [new branch] gh/yanbing-j/26/head -> origin/gh/yanbing-j/26/head 2025-12-04T08:53:58.7782996Z * [new branch] gh/yanbing-j/26/orig -> origin/gh/yanbing-j/26/orig 2025-12-04T08:53:58.7783076Z * [new branch] gh/yang-yu-hang/1/base -> origin/gh/yang-yu-hang/1/base 2025-12-04T08:53:58.7783152Z * [new branch] gh/yang-yu-hang/1/head -> origin/gh/yang-yu-hang/1/head 2025-12-04T08:53:58.7783226Z * [new branch] gh/yang-yu-hang/1/orig -> origin/gh/yang-yu-hang/1/orig 2025-12-04T08:53:58.7783301Z * [new branch] gh/yang-yu-hang/2/base -> origin/gh/yang-yu-hang/2/base 2025-12-04T08:53:58.7783375Z * [new branch] gh/yang-yu-hang/2/head -> origin/gh/yang-yu-hang/2/head 2025-12-04T08:53:58.7783452Z * [new branch] gh/yang-yu-hang/2/orig -> origin/gh/yang-yu-hang/2/orig 2025-12-04T08:53:58.7783525Z * [new branch] gh/yang-yu-hang/3/base -> origin/gh/yang-yu-hang/3/base 2025-12-04T08:53:58.7783597Z * [new branch] gh/yang-yu-hang/3/head -> origin/gh/yang-yu-hang/3/head 2025-12-04T08:53:58.7783672Z * [new branch] gh/yang-yu-hang/3/orig -> origin/gh/yang-yu-hang/3/orig 2025-12-04T08:53:58.7783748Z * [new branch] gh/yangw-dev/12/base -> origin/gh/yangw-dev/12/base 2025-12-04T08:53:58.7783820Z * [new branch] gh/yangw-dev/12/head -> origin/gh/yangw-dev/12/head 2025-12-04T08:53:58.7783893Z * [new branch] gh/yangw-dev/12/orig -> origin/gh/yangw-dev/12/orig 2025-12-04T08:53:58.7783964Z * [new branch] gh/yangw-dev/13/base -> origin/gh/yangw-dev/13/base 2025-12-04T08:53:58.7784034Z * [new branch] gh/yangw-dev/13/head -> origin/gh/yangw-dev/13/head 2025-12-04T08:53:58.7784130Z * [new branch] gh/yangw-dev/13/orig -> origin/gh/yangw-dev/13/orig 2025-12-04T08:53:58.7784201Z * [new branch] gh/yangw-dev/14/base -> origin/gh/yangw-dev/14/base 2025-12-04T08:53:58.7784271Z * [new branch] gh/yangw-dev/14/head -> origin/gh/yangw-dev/14/head 2025-12-04T08:53:58.7784344Z * [new branch] gh/yangw-dev/14/orig -> origin/gh/yangw-dev/14/orig 2025-12-04T08:53:58.7784415Z * [new branch] gh/yangw-dev/15/base -> origin/gh/yangw-dev/15/base 2025-12-04T08:53:58.7784485Z * [new branch] gh/yangw-dev/15/head -> origin/gh/yangw-dev/15/head 2025-12-04T08:53:58.7784559Z * [new branch] gh/yangw-dev/15/orig -> origin/gh/yangw-dev/15/orig 2025-12-04T08:53:58.7784630Z * [new branch] gh/yangw-dev/19/base -> origin/gh/yangw-dev/19/base 2025-12-04T08:53:58.7784700Z * [new branch] gh/yangw-dev/19/head -> origin/gh/yangw-dev/19/head 2025-12-04T08:53:58.7784775Z * [new branch] gh/yangw-dev/19/orig -> origin/gh/yangw-dev/19/orig 2025-12-04T08:53:58.7784845Z * [new branch] gh/yangw-dev/26/base -> origin/gh/yangw-dev/26/base 2025-12-04T08:53:58.7784916Z * [new branch] gh/yangw-dev/26/head -> origin/gh/yangw-dev/26/head 2025-12-04T08:53:58.7785027Z * [new branch] gh/yangw-dev/26/orig -> origin/gh/yangw-dev/26/orig 2025-12-04T08:53:58.7785098Z * [new branch] gh/yangw-dev/27/base -> origin/gh/yangw-dev/27/base 2025-12-04T08:53:58.7785170Z * [new branch] gh/yangw-dev/27/head -> origin/gh/yangw-dev/27/head 2025-12-04T08:53:58.7785243Z * [new branch] gh/yangw-dev/27/orig -> origin/gh/yangw-dev/27/orig 2025-12-04T08:53:58.7785313Z * [new branch] gh/ydwu4/292/base -> origin/gh/ydwu4/292/base 2025-12-04T08:53:58.7785383Z * [new branch] gh/ydwu4/292/head -> origin/gh/ydwu4/292/head 2025-12-04T08:53:58.7785453Z * [new branch] gh/ydwu4/292/orig -> origin/gh/ydwu4/292/orig 2025-12-04T08:53:58.7785519Z * [new branch] gh/ydwu4/294/base -> origin/gh/ydwu4/294/base 2025-12-04T08:53:58.7785588Z * [new branch] gh/ydwu4/294/head -> origin/gh/ydwu4/294/head 2025-12-04T08:53:58.7785655Z * [new branch] gh/ydwu4/294/orig -> origin/gh/ydwu4/294/orig 2025-12-04T08:53:58.7785721Z * [new branch] gh/ydwu4/295/base -> origin/gh/ydwu4/295/base 2025-12-04T08:53:58.7785788Z * [new branch] gh/ydwu4/295/head -> origin/gh/ydwu4/295/head 2025-12-04T08:53:58.7785854Z * [new branch] gh/ydwu4/295/orig -> origin/gh/ydwu4/295/orig 2025-12-04T08:53:58.7785920Z * [new branch] gh/ydwu4/296/base -> origin/gh/ydwu4/296/base 2025-12-04T08:53:58.7785988Z * [new branch] gh/ydwu4/296/head -> origin/gh/ydwu4/296/head 2025-12-04T08:53:58.7786055Z * [new branch] gh/ydwu4/296/orig -> origin/gh/ydwu4/296/orig 2025-12-04T08:53:58.7786121Z * [new branch] gh/ydwu4/306/base -> origin/gh/ydwu4/306/base 2025-12-04T08:53:58.7786189Z * [new branch] gh/ydwu4/306/head -> origin/gh/ydwu4/306/head 2025-12-04T08:53:58.7786256Z * [new branch] gh/ydwu4/306/orig -> origin/gh/ydwu4/306/orig 2025-12-04T08:53:58.7786323Z * [new branch] gh/ydwu4/312/base -> origin/gh/ydwu4/312/base 2025-12-04T08:53:58.7786391Z * [new branch] gh/ydwu4/312/head -> origin/gh/ydwu4/312/head 2025-12-04T08:53:58.7786458Z * [new branch] gh/ydwu4/312/orig -> origin/gh/ydwu4/312/orig 2025-12-04T08:53:58.7786525Z * [new branch] gh/ydwu4/322/base -> origin/gh/ydwu4/322/base 2025-12-04T08:53:58.7786593Z * [new branch] gh/ydwu4/322/head -> origin/gh/ydwu4/322/head 2025-12-04T08:53:58.7786694Z * [new branch] gh/ydwu4/322/orig -> origin/gh/ydwu4/322/orig 2025-12-04T08:53:58.7786810Z * [new branch] gh/ydwu4/327/base -> origin/gh/ydwu4/327/base 2025-12-04T08:53:58.7786880Z * [new branch] gh/ydwu4/327/head -> origin/gh/ydwu4/327/head 2025-12-04T08:53:58.7786947Z * [new branch] gh/ydwu4/327/orig -> origin/gh/ydwu4/327/orig 2025-12-04T08:53:58.7787015Z * [new branch] gh/ydwu4/328/base -> origin/gh/ydwu4/328/base 2025-12-04T08:53:58.7787081Z * [new branch] gh/ydwu4/328/head -> origin/gh/ydwu4/328/head 2025-12-04T08:53:58.7787148Z * [new branch] gh/ydwu4/328/orig -> origin/gh/ydwu4/328/orig 2025-12-04T08:53:58.7787216Z * [new branch] gh/ydwu4/329/base -> origin/gh/ydwu4/329/base 2025-12-04T08:53:58.7787283Z * [new branch] gh/ydwu4/329/head -> origin/gh/ydwu4/329/head 2025-12-04T08:53:58.7787352Z * [new branch] gh/ydwu4/329/orig -> origin/gh/ydwu4/329/orig 2025-12-04T08:53:58.7787421Z * [new branch] gh/ydwu4/330/base -> origin/gh/ydwu4/330/base 2025-12-04T08:53:58.7787487Z * [new branch] gh/ydwu4/330/head -> origin/gh/ydwu4/330/head 2025-12-04T08:53:58.7787603Z * [new branch] gh/ydwu4/330/orig -> origin/gh/ydwu4/330/orig 2025-12-04T08:53:58.7787671Z * [new branch] gh/ydwu4/331/base -> origin/gh/ydwu4/331/base 2025-12-04T08:53:58.7787738Z * [new branch] gh/ydwu4/331/head -> origin/gh/ydwu4/331/head 2025-12-04T08:53:58.7787804Z * [new branch] gh/ydwu4/331/orig -> origin/gh/ydwu4/331/orig 2025-12-04T08:53:58.7787872Z * [new branch] gh/ydwu4/332/base -> origin/gh/ydwu4/332/base 2025-12-04T08:53:58.7787938Z * [new branch] gh/ydwu4/332/head -> origin/gh/ydwu4/332/head 2025-12-04T08:53:58.7788007Z * [new branch] gh/ydwu4/332/orig -> origin/gh/ydwu4/332/orig 2025-12-04T08:53:58.7788075Z * [new branch] gh/ydwu4/333/base -> origin/gh/ydwu4/333/base 2025-12-04T08:53:58.7788141Z * [new branch] gh/ydwu4/333/head -> origin/gh/ydwu4/333/head 2025-12-04T08:53:58.7788207Z * [new branch] gh/ydwu4/333/orig -> origin/gh/ydwu4/333/orig 2025-12-04T08:53:58.7788277Z * [new branch] gh/ydwu4/334/base -> origin/gh/ydwu4/334/base 2025-12-04T08:53:58.7788344Z * [new branch] gh/ydwu4/334/head -> origin/gh/ydwu4/334/head 2025-12-04T08:53:58.7788411Z * [new branch] gh/ydwu4/334/orig -> origin/gh/ydwu4/334/orig 2025-12-04T08:53:58.7788479Z * [new branch] gh/ydwu4/335/base -> origin/gh/ydwu4/335/base 2025-12-04T08:53:58.7788545Z * [new branch] gh/ydwu4/335/head -> origin/gh/ydwu4/335/head 2025-12-04T08:53:58.7788613Z * [new branch] gh/ydwu4/335/orig -> origin/gh/ydwu4/335/orig 2025-12-04T08:53:58.7788683Z * [new branch] gh/ydwu4/337/base -> origin/gh/ydwu4/337/base 2025-12-04T08:53:58.7788750Z * [new branch] gh/ydwu4/337/head -> origin/gh/ydwu4/337/head 2025-12-04T08:53:58.7788818Z * [new branch] gh/ydwu4/337/orig -> origin/gh/ydwu4/337/orig 2025-12-04T08:53:58.7791587Z * [new branch] gh/ydwu4/339/base -> origin/gh/ydwu4/339/base 2025-12-04T08:53:58.7791663Z * [new branch] gh/ydwu4/339/head -> origin/gh/ydwu4/339/head 2025-12-04T08:53:58.7791732Z * [new branch] gh/ydwu4/339/orig -> origin/gh/ydwu4/339/orig 2025-12-04T08:53:58.7791798Z * [new branch] gh/yf225/133/base -> origin/gh/yf225/133/base 2025-12-04T08:53:58.7791864Z * [new branch] gh/yf225/133/head -> origin/gh/yf225/133/head 2025-12-04T08:53:58.7791990Z * [new branch] gh/yf225/93/base -> origin/gh/yf225/93/base 2025-12-04T08:53:58.7792057Z * [new branch] gh/yf225/93/head -> origin/gh/yf225/93/head 2025-12-04T08:53:58.7792133Z * [new branch] gh/yifuwang/152/base -> origin/gh/yifuwang/152/base 2025-12-04T08:53:58.7792208Z * [new branch] gh/yifuwang/152/head -> origin/gh/yifuwang/152/head 2025-12-04T08:53:58.7792280Z * [new branch] gh/yifuwang/152/orig -> origin/gh/yifuwang/152/orig 2025-12-04T08:53:58.7792352Z * [new branch] gh/yifuwang/195/base -> origin/gh/yifuwang/195/base 2025-12-04T08:53:58.7792426Z * [new branch] gh/yifuwang/195/head -> origin/gh/yifuwang/195/head 2025-12-04T08:53:58.7792497Z * [new branch] gh/yifuwang/195/orig -> origin/gh/yifuwang/195/orig 2025-12-04T08:53:58.7792570Z * [new branch] gh/yiming0416/1/base -> origin/gh/yiming0416/1/base 2025-12-04T08:53:58.7792645Z * [new branch] gh/yiming0416/1/head -> origin/gh/yiming0416/1/head 2025-12-04T08:53:58.7792716Z * [new branch] gh/yiming0416/2/base -> origin/gh/yiming0416/2/base 2025-12-04T08:53:58.7792787Z * [new branch] gh/yiming0416/2/head -> origin/gh/yiming0416/2/head 2025-12-04T08:53:58.7792862Z * [new branch] gh/yushangdi/1/base -> origin/gh/yushangdi/1/base 2025-12-04T08:53:58.7792967Z * [new branch] gh/yushangdi/1/head -> origin/gh/yushangdi/1/head 2025-12-04T08:53:58.7793041Z * [new branch] gh/yushangdi/10/base -> origin/gh/yushangdi/10/base 2025-12-04T08:53:58.7793113Z * [new branch] gh/yushangdi/10/head -> origin/gh/yushangdi/10/head 2025-12-04T08:53:58.7793186Z * [new branch] gh/yushangdi/10/orig -> origin/gh/yushangdi/10/orig 2025-12-04T08:53:58.7793258Z * [new branch] gh/yushangdi/11/base -> origin/gh/yushangdi/11/base 2025-12-04T08:53:58.7793333Z * [new branch] gh/yushangdi/11/head -> origin/gh/yushangdi/11/head 2025-12-04T08:53:58.7793404Z * [new branch] gh/yushangdi/11/orig -> origin/gh/yushangdi/11/orig 2025-12-04T08:53:58.7793475Z * [new branch] gh/yushangdi/2/base -> origin/gh/yushangdi/2/base 2025-12-04T08:53:58.7793547Z * [new branch] gh/yushangdi/2/head -> origin/gh/yushangdi/2/head 2025-12-04T08:53:58.7793621Z * [new branch] gh/yushangdi/7/base -> origin/gh/yushangdi/7/base 2025-12-04T08:53:58.7793693Z * [new branch] gh/yushangdi/7/head -> origin/gh/yushangdi/7/head 2025-12-04T08:53:58.7793764Z * [new branch] gh/yushangdi/7/orig -> origin/gh/yushangdi/7/orig 2025-12-04T08:53:58.7793834Z * [new branch] gh/yushangdi/8/base -> origin/gh/yushangdi/8/base 2025-12-04T08:53:58.7793905Z * [new branch] gh/yushangdi/8/head -> origin/gh/yushangdi/8/head 2025-12-04T08:53:58.7793978Z * [new branch] gh/yushangdi/8/orig -> origin/gh/yushangdi/8/orig 2025-12-04T08:53:58.7794048Z * [new branch] gh/yushangdi/9/base -> origin/gh/yushangdi/9/base 2025-12-04T08:53:58.7794119Z * [new branch] gh/yushangdi/9/head -> origin/gh/yushangdi/9/head 2025-12-04T08:53:58.7794189Z * [new branch] gh/yushangdi/9/orig -> origin/gh/yushangdi/9/orig 2025-12-04T08:53:58.7794262Z * [new branch] gh/zklaus/19/base -> origin/gh/zklaus/19/base 2025-12-04T08:53:58.7794332Z * [new branch] gh/zklaus/19/head -> origin/gh/zklaus/19/head 2025-12-04T08:53:58.7794399Z * [new branch] gh/zklaus/19/orig -> origin/gh/zklaus/19/orig 2025-12-04T08:53:58.7794466Z * [new branch] gh/zklaus/20/base -> origin/gh/zklaus/20/base 2025-12-04T08:53:58.7794534Z * [new branch] gh/zklaus/20/head -> origin/gh/zklaus/20/head 2025-12-04T08:53:58.7794626Z * [new branch] gh/zklaus/20/orig -> origin/gh/zklaus/20/orig 2025-12-04T08:53:58.7794693Z * [new branch] gh/zklaus/21/base -> origin/gh/zklaus/21/base 2025-12-04T08:53:58.7794761Z * [new branch] gh/zklaus/21/head -> origin/gh/zklaus/21/head 2025-12-04T08:53:58.7794829Z * [new branch] gh/zklaus/21/orig -> origin/gh/zklaus/21/orig 2025-12-04T08:53:58.7794897Z * [new branch] gh/zklaus/22/base -> origin/gh/zklaus/22/base 2025-12-04T08:53:58.7794965Z * [new branch] gh/zklaus/22/head -> origin/gh/zklaus/22/head 2025-12-04T08:53:58.7795032Z * [new branch] gh/zklaus/22/orig -> origin/gh/zklaus/22/orig 2025-12-04T08:53:58.7795098Z * [new branch] gh/zklaus/23/base -> origin/gh/zklaus/23/base 2025-12-04T08:53:58.7795167Z * [new branch] gh/zklaus/23/head -> origin/gh/zklaus/23/head 2025-12-04T08:53:58.7795236Z * [new branch] gh/zklaus/23/orig -> origin/gh/zklaus/23/orig 2025-12-04T08:53:58.7795302Z * [new branch] gh/zklaus/24/base -> origin/gh/zklaus/24/base 2025-12-04T08:53:58.7795370Z * [new branch] gh/zklaus/24/head -> origin/gh/zklaus/24/head 2025-12-04T08:53:58.7795436Z * [new branch] gh/zklaus/24/orig -> origin/gh/zklaus/24/orig 2025-12-04T08:53:58.7795533Z * [new branch] gh/zou3519/1197/base -> origin/gh/zou3519/1197/base 2025-12-04T08:53:58.7795606Z * [new branch] gh/zou3519/1197/head -> origin/gh/zou3519/1197/head 2025-12-04T08:53:58.7795676Z * [new branch] gh/zou3519/1197/orig -> origin/gh/zou3519/1197/orig 2025-12-04T08:53:58.7795745Z * [new branch] gh/zou3519/1199/base -> origin/gh/zou3519/1199/base 2025-12-04T08:53:58.7795816Z * [new branch] gh/zou3519/1199/head -> origin/gh/zou3519/1199/head 2025-12-04T08:53:58.7795886Z * [new branch] gh/zou3519/1199/orig -> origin/gh/zou3519/1199/orig 2025-12-04T08:53:58.7795954Z * [new branch] gh/zou3519/1200/base -> origin/gh/zou3519/1200/base 2025-12-04T08:53:58.7796023Z * [new branch] gh/zou3519/1200/head -> origin/gh/zou3519/1200/head 2025-12-04T08:53:58.7796091Z * [new branch] gh/zou3519/1200/orig -> origin/gh/zou3519/1200/orig 2025-12-04T08:53:58.7796162Z * [new branch] gh/zou3519/1201/base -> origin/gh/zou3519/1201/base 2025-12-04T08:53:58.7796230Z * [new branch] gh/zou3519/1201/head -> origin/gh/zou3519/1201/head 2025-12-04T08:53:58.7796299Z * [new branch] gh/zou3519/1201/orig -> origin/gh/zou3519/1201/orig 2025-12-04T08:53:58.7796368Z * [new branch] gh/zou3519/1202/base -> origin/gh/zou3519/1202/base 2025-12-04T08:53:58.7796436Z * [new branch] gh/zou3519/1202/head -> origin/gh/zou3519/1202/head 2025-12-04T08:53:58.7796507Z * [new branch] gh/zou3519/1202/orig -> origin/gh/zou3519/1202/orig 2025-12-04T08:53:58.7796578Z * [new branch] gh/zpcore/1/base -> origin/gh/zpcore/1/base 2025-12-04T08:53:58.7796646Z * [new branch] gh/zpcore/1/head -> origin/gh/zpcore/1/head 2025-12-04T08:53:58.7796714Z * [new branch] gh/zpcore/11/base -> origin/gh/zpcore/11/base 2025-12-04T08:53:58.7796831Z * [new branch] gh/zpcore/11/head -> origin/gh/zpcore/11/head 2025-12-04T08:53:58.7796902Z * [new branch] gh/zpcore/11/orig -> origin/gh/zpcore/11/orig 2025-12-04T08:53:58.7796970Z * [new branch] gh/zpcore/12/base -> origin/gh/zpcore/12/base 2025-12-04T08:53:58.7797038Z * [new branch] gh/zpcore/12/head -> origin/gh/zpcore/12/head 2025-12-04T08:53:58.7797106Z * [new branch] gh/zpcore/12/orig -> origin/gh/zpcore/12/orig 2025-12-04T08:53:58.7797209Z * [new branch] gh/zpcore/13/base -> origin/gh/zpcore/13/base 2025-12-04T08:53:58.7797277Z * [new branch] gh/zpcore/13/head -> origin/gh/zpcore/13/head 2025-12-04T08:53:58.7797344Z * [new branch] gh/zpcore/13/orig -> origin/gh/zpcore/13/orig 2025-12-04T08:53:58.7797411Z * [new branch] gh/zpcore/14/base -> origin/gh/zpcore/14/base 2025-12-04T08:53:58.7797480Z * [new branch] gh/zpcore/14/head -> origin/gh/zpcore/14/head 2025-12-04T08:53:58.7797548Z * [new branch] gh/zpcore/14/orig -> origin/gh/zpcore/14/orig 2025-12-04T08:53:58.7797616Z * [new branch] gh/zpcore/15/base -> origin/gh/zpcore/15/base 2025-12-04T08:53:58.7797684Z * [new branch] gh/zpcore/15/head -> origin/gh/zpcore/15/head 2025-12-04T08:53:58.7797752Z * [new branch] gh/zpcore/15/orig -> origin/gh/zpcore/15/orig 2025-12-04T08:53:58.7797820Z * [new branch] gh/zpcore/2/base -> origin/gh/zpcore/2/base 2025-12-04T08:53:58.7797890Z * [new branch] gh/zpcore/2/head -> origin/gh/zpcore/2/head 2025-12-04T08:53:58.7797956Z * [new branch] gh/zpcore/21/base -> origin/gh/zpcore/21/base 2025-12-04T08:53:58.7798024Z * [new branch] gh/zpcore/21/head -> origin/gh/zpcore/21/head 2025-12-04T08:53:58.7798139Z * [new branch] gh/zpcore/21/orig -> origin/gh/zpcore/21/orig 2025-12-04T08:53:58.7798207Z * [new branch] gh/zpcore/22/base -> origin/gh/zpcore/22/base 2025-12-04T08:53:58.7798275Z * [new branch] gh/zpcore/22/head -> origin/gh/zpcore/22/head 2025-12-04T08:53:58.7798342Z * [new branch] gh/zpcore/22/orig -> origin/gh/zpcore/22/orig 2025-12-04T08:53:58.7798409Z * [new branch] gh/zpcore/23/base -> origin/gh/zpcore/23/base 2025-12-04T08:53:58.7798476Z * [new branch] gh/zpcore/23/head -> origin/gh/zpcore/23/head 2025-12-04T08:53:58.7798545Z * [new branch] gh/zpcore/23/orig -> origin/gh/zpcore/23/orig 2025-12-04T08:53:58.7798612Z * [new branch] gh/zpcore/24/base -> origin/gh/zpcore/24/base 2025-12-04T08:53:58.7798679Z * [new branch] gh/zpcore/24/head -> origin/gh/zpcore/24/head 2025-12-04T08:53:58.7798748Z * [new branch] gh/zpcore/24/orig -> origin/gh/zpcore/24/orig 2025-12-04T08:53:58.7798816Z * [new branch] gh/zpcore/25/base -> origin/gh/zpcore/25/base 2025-12-04T08:53:58.7798884Z * [new branch] gh/zpcore/25/head -> origin/gh/zpcore/25/head 2025-12-04T08:53:58.7798951Z * [new branch] gh/zpcore/25/orig -> origin/gh/zpcore/25/orig 2025-12-04T08:53:58.7799018Z * [new branch] gh/zpcore/26/base -> origin/gh/zpcore/26/base 2025-12-04T08:53:58.7799086Z * [new branch] gh/zpcore/26/head -> origin/gh/zpcore/26/head 2025-12-04T08:53:58.7799154Z * [new branch] gh/zpcore/26/orig -> origin/gh/zpcore/26/orig 2025-12-04T08:53:58.7799220Z * [new branch] gh/zpcore/27/base -> origin/gh/zpcore/27/base 2025-12-04T08:53:58.7799287Z * [new branch] gh/zpcore/27/head -> origin/gh/zpcore/27/head 2025-12-04T08:53:58.7799354Z * [new branch] gh/zpcore/27/orig -> origin/gh/zpcore/27/orig 2025-12-04T08:53:58.7799422Z * [new branch] gh/zpcore/28/base -> origin/gh/zpcore/28/base 2025-12-04T08:53:58.7799489Z * [new branch] gh/zpcore/28/head -> origin/gh/zpcore/28/head 2025-12-04T08:53:58.7799555Z * [new branch] gh/zpcore/28/orig -> origin/gh/zpcore/28/orig 2025-12-04T08:53:58.7799623Z * [new branch] gh/zpcore/3/base -> origin/gh/zpcore/3/base 2025-12-04T08:53:58.7799689Z * [new branch] gh/zpcore/3/head -> origin/gh/zpcore/3/head 2025-12-04T08:53:58.7799780Z * [new branch] gh/zpcore/4/base -> origin/gh/zpcore/4/base 2025-12-04T08:53:58.7799850Z * [new branch] gh/zpcore/4/head -> origin/gh/zpcore/4/head 2025-12-04T08:53:58.7799918Z * [new branch] gh/zpcore/5/base -> origin/gh/zpcore/5/base 2025-12-04T08:53:58.7799985Z * [new branch] gh/zpcore/5/head -> origin/gh/zpcore/5/head 2025-12-04T08:53:58.7800054Z * [new branch] gh/zpcore/6/base -> origin/gh/zpcore/6/base 2025-12-04T08:53:58.7800123Z * [new branch] gh/zpcore/6/head -> origin/gh/zpcore/6/head 2025-12-04T08:53:58.7800190Z * [new branch] gh/zpcore/7/base -> origin/gh/zpcore/7/base 2025-12-04T08:53:58.7800258Z * [new branch] gh/zpcore/7/head -> origin/gh/zpcore/7/head 2025-12-04T08:53:58.7800325Z * [new branch] gh/zpcore/8/base -> origin/gh/zpcore/8/base 2025-12-04T08:53:58.7800392Z * [new branch] gh/zpcore/8/head -> origin/gh/zpcore/8/head 2025-12-04T08:53:58.7800463Z * [new branch] google-main -> origin/google-main 2025-12-04T08:53:58.7800550Z * [new branch] guangyey/external_stream -> origin/guangyey/external_stream 2025-12-04T08:53:58.7800623Z * [new branch] guangyey/test_2025 -> origin/guangyey/test_2025 2025-12-04T08:53:58.7800789Z * [new branch] guilhermeleobas/cherry-pick-55d87d9dfd9 -> origin/guilhermeleobas/cherry-pick-55d87d9dfd9 2025-12-04T08:53:58.7800908Z * [new branch] hameerabbasi/complex_tensor_subclass -> origin/hameerabbasi/complex_tensor_subclass 2025-12-04T08:53:58.7801049Z * [new branch] hameerabbasi/fix-ctensor-gradcheck-tests -> origin/hameerabbasi/fix-ctensor-gradcheck-tests 2025-12-04T08:53:58.7801158Z * [new branch] hameerabbasi/gradcheck-allclose -> origin/hameerabbasi/gradcheck-allclose 2025-12-04T08:53:58.7801224Z * [new branch] hc_baseline -> origin/hc_baseline 2025-12-04T08:53:58.7801287Z * [new branch] hhh_rand -> origin/hhh_rand 2025-12-04T08:53:58.7801349Z * [new branch] huba/f1 -> origin/huba/f1 2025-12-04T08:53:58.7801538Z * [new branch] increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test -> origin/increase-timeout-linux-jammy-cuda12_8-py3_10-gcc11-test 2025-12-04T08:53:58.7801602Z * [new branch] inlining -> origin/inlining 2025-12-04T08:53:58.7801675Z * [new branch] inlining-ezyang -> origin/inlining-ezyang 2025-12-04T08:53:58.7801761Z * [new branch] install-torchao-0.13.0 -> origin/install-torchao-0.13.0 2025-12-04T08:53:58.7801944Z * [new branch] instrument-trunk-pull-linux-with-job-test-filters -> origin/instrument-trunk-pull-linux-with-job-test-filters 2025-12-04T08:53:58.7802017Z * [new branch] invoke-subgraph -> origin/invoke-subgraph 2025-12-04T08:53:58.7802083Z * [new branch] issue#58739 -> origin/issue#58739 2025-12-04T08:53:58.7802164Z * [new branch] jainapurva-patch-1 -> origin/jainapurva-patch-1 2025-12-04T08:53:58.7802226Z * [new branch] jathu/o3 -> origin/jathu/o3 2025-12-04T08:53:58.7802289Z * [new branch] jathu/sve -> origin/jathu/sve 2025-12-04T08:53:58.7802414Z * [new branch] jcaip/test-cusparselt-version-0.6.2 -> origin/jcaip/test-cusparselt-version-0.6.2 2025-12-04T08:53:58.7802518Z * [new branch] jcaip/update-cusparselt-0.6.2 -> origin/jcaip/update-cusparselt-0.6.2 2025-12-04T08:53:58.7802632Z * [new branch] jiannanWang/memorysnapshot_filter -> origin/jiannanWang/memorysnapshot_filter 2025-12-04T08:53:58.7802742Z * [new branch] jiannanWang/profilerstepwarning -> origin/jiannanWang/profilerstepwarning 2025-12-04T08:53:58.7802854Z * [new branch] jithunnair-amd-patch-1 -> origin/jithunnair-amd-patch-1 2025-12-04T08:53:58.7802940Z * [new branch] jithunnair-amd-patch-10 -> origin/jithunnair-amd-patch-10 2025-12-04T08:53:58.7803021Z * [new branch] jithunnair-amd-patch-2 -> origin/jithunnair-amd-patch-2 2025-12-04T08:53:58.7803103Z * [new branch] jithunnair-amd-patch-3 -> origin/jithunnair-amd-patch-3 2025-12-04T08:53:58.7803184Z * [new branch] jithunnair-amd-patch-4 -> origin/jithunnair-amd-patch-4 2025-12-04T08:53:58.7803262Z * [new branch] jithunnair-amd-patch-5 -> origin/jithunnair-amd-patch-5 2025-12-04T08:53:58.7803341Z * [new branch] jithunnair-amd-patch-6 -> origin/jithunnair-amd-patch-6 2025-12-04T08:53:58.7803424Z * [new branch] jithunnair-amd-patch-7 -> origin/jithunnair-amd-patch-7 2025-12-04T08:53:58.7803502Z * [new branch] jithunnair-amd-patch-8 -> origin/jithunnair-amd-patch-8 2025-12-04T08:53:58.7803582Z * [new branch] jithunnair-amd-patch-9 -> origin/jithunnair-amd-patch-9 2025-12-04T08:53:58.7803660Z * [new branch] justinchu/native-qdq -> origin/justinchu/native-qdq 2025-12-04T08:53:58.7803732Z * [new branch] kainan666/xlf_debug -> origin/kainan666/xlf_debug 2025-12-04T08:53:58.7803820Z * [new branch] kainan_test -> origin/kainan_test 2025-12-04T08:53:58.7803900Z * [new branch] larryliu0820-patch-1 -> origin/larryliu0820-patch-1 2025-12-04T08:53:58.7804007Z * [new branch] leslie/test_group_gemm_epilogues -> origin/leslie/test_group_gemm_epilogues 2025-12-04T08:53:58.7804111Z * [new branch] lessw2020/fix_cutlass_cache_error -> origin/lessw2020/fix_cutlass_cache_error 2025-12-04T08:53:58.7804189Z * [new branch] liaoxuan/shm_all_reduce -> origin/liaoxuan/shm_all_reduce 2025-12-04T08:53:58.7804294Z * [new branch] liaoxuan/test_fa_disable_softmax -> origin/liaoxuan/test_fa_disable_softmax 2025-12-04T08:53:58.7804374Z * [new branch] liaoxuan/test_int8_sdpa -> origin/liaoxuan/test_int8_sdpa 2025-12-04T08:53:58.7804444Z * [new branch] llama4-stable -> origin/llama4-stable 2025-12-04T08:53:58.7804511Z * [new branch] lts/release/1.8 -> origin/lts/release/1.8 2025-12-04T08:53:58.7804587Z * [new branch] lucaskabela/#94773 -> origin/lucaskabela/#94773 2025-12-04T08:53:58.7804663Z * [new branch] lucaskabela/fix_164876 -> origin/lucaskabela/fix_164876 2025-12-04T08:53:58.7804748Z * [new branch] lucaskabela/flop_counter -> origin/lucaskabela/flop_counter 2025-12-04T08:53:58.7804844Z * [new branch] lucaskabela/func_under_decomp -> origin/lucaskabela/func_under_decomp 2025-12-04T08:53:58.7804949Z * [new branch] lucaskabela/functional_in_dynamo -> origin/lucaskabela/functional_in_dynamo 2025-12-04T08:53:58.7805077Z * [new branch] lucaskabela/install_params_as_graph_attr -> origin/lucaskabela/install_params_as_graph_attr 2025-12-04T08:53:58.7805194Z * [new branch] lucaskabela/parameters_as_graph_attr -> origin/lucaskabela/parameters_as_graph_attr 2025-12-04T08:53:58.7805328Z * [new branch] lucaskabela/remove_aot_dispatcher_metadata -> origin/lucaskabela/remove_aot_dispatcher_metadata 2025-12-04T08:53:58.7805407Z * [new branch] lucaskabela/rnn_decomp -> origin/lucaskabela/rnn_decomp 2025-12-04T08:53:58.7805502Z * [new branch] lucaskabela/typing_backends -> origin/lucaskabela/typing_backends 2025-12-04T08:53:58.7805602Z * [new branch] lucaskabela/typing_ctx_manager -> origin/lucaskabela/typing_ctx_manager 2025-12-04T08:53:58.7805696Z * [new branch] lucaskabela/typing_nn_module -> origin/lucaskabela/typing_nn_module 2025-12-04T08:53:58.7805823Z * [new branch] lucaskabela/typing_user_defined -> origin/lucaskabela/typing_user_defined 2025-12-04T08:53:58.7805919Z * [new branch] lucaskabela/typing_variables -> origin/lucaskabela/typing_variables 2025-12-04T08:53:58.7806032Z * [new branch] lucaskabela/typing_variables_dicts -> origin/lucaskabela/typing_variables_dicts 2025-12-04T08:53:58.7806157Z * [new branch] lucaskabela/typing_variables_functions -> origin/lucaskabela/typing_variables_functions 2025-12-04T08:53:58.7806266Z * [new branch] lucaskabela/typing_variables_lists -> origin/lucaskabela/typing_variables_lists 2025-12-04T08:53:58.7806340Z * [new branch] lw/torch_box_by_ref -> origin/lw/torch_box_by_ref 2025-12-04T08:53:58.7806402Z * [new branch] main -> origin/main 2025-12-04T08:53:58.7806472Z * [new branch] malfet-patch-1 -> origin/malfet-patch-1 2025-12-04T08:53:58.7806543Z * [new branch] malfet-patch-2 -> origin/malfet-patch-2 2025-12-04T08:53:58.7806610Z * [new branch] malfet-patch-3 -> origin/malfet-patch-3 2025-12-04T08:53:58.7806677Z * [new branch] malfet-patch-4 -> origin/malfet-patch-4 2025-12-04T08:53:58.7806786Z * [new branch] malfet-patch-5 -> origin/malfet-patch-5 2025-12-04T08:53:58.7806897Z * [new branch] malfet-patch-6 -> origin/malfet-patch-6 2025-12-04T08:53:58.7806964Z * [new branch] malfet-patch-7 -> origin/malfet-patch-7 2025-12-04T08:53:58.7807031Z * [new branch] malfet-patch-8 -> origin/malfet-patch-8 2025-12-04T08:53:58.7807106Z * [new branch] malfet/add-3.14-ci -> origin/malfet/add-3.14-ci 2025-12-04T08:53:58.7807268Z * [new branch] malfet/be-do-not-make-typos-in-build-artifacts -> origin/malfet/be-do-not-make-typos-in-build-artifacts 2025-12-04T08:53:58.7807439Z * [new branch] malfet/be-move-more-settings-to-checkout-pytorch -> origin/malfet/be-move-more-settings-to-checkout-pytorch 2025-12-04T08:53:58.7807566Z * [new branch] malfet/be-remove-misisng-neon-headers -> origin/malfet/be-remove-misisng-neon-headers 2025-12-04T08:53:58.7807666Z * [new branch] malfet/mps-implement-col2im -> origin/malfet/mps-implement-col2im 2025-12-04T08:53:58.7807783Z * [new branch] manuel/aoti_metal_shimify-thread_safe -> origin/manuel/aoti_metal_shimify-thread_safe 2025-12-04T08:53:58.7807877Z * [new branch] manuel/inductor_link_openmp -> origin/manuel/inductor_link_openmp 2025-12-04T08:53:58.7807954Z * [new branch] masnesral/metaconda -> origin/masnesral/metaconda 2025-12-04T08:53:58.7808030Z * [new branch] mem_profiler_flaky_fix -> origin/mem_profiler_flaky_fix 2025-12-04T08:53:58.7808109Z * [new branch] mem_profiler_stack_trace -> origin/mem_profiler_stack_trace 2025-12-04T08:53:58.7808192Z * [new branch] memory_profiler_stack -> origin/memory_profiler_stack 2025-12-04T08:53:58.7808266Z * [new branch] metascroy-patch-1 -> origin/metascroy-patch-1 2025-12-04T08:53:58.7808330Z * [new branch] mingw_posix -> origin/mingw_posix 2025-12-04T08:53:58.7808407Z * [new branch] mlazos/S429861-debug -> origin/mlazos/S429861-debug 2025-12-04T08:53:58.7808474Z * [new branch] mlazos/aa -> origin/mlazos/aa 2025-12-04T08:53:58.7808538Z * [new branch] mlazos/acts -> origin/mlazos/acts 2025-12-04T08:53:58.7808612Z * [new branch] mlazos/arg-renames -> origin/mlazos/arg-renames 2025-12-04T08:53:58.7808691Z * [new branch] mlazos/bad-cudagraphs -> origin/mlazos/bad-cudagraphs 2025-12-04T08:53:58.7808792Z * [new branch] mlazos/baseline-graph-breaks -> origin/mlazos/baseline-graph-breaks 2025-12-04T08:53:58.7808905Z * [new branch] mlazos/beta-tensor -> origin/mlazos/beta-tensor 2025-12-04T08:53:58.7808971Z * [new branch] mlazos/buffers -> origin/mlazos/buffers 2025-12-04T08:53:58.7809039Z * [new branch] mlazos/buffers2 -> origin/mlazos/buffers2 2025-12-04T08:53:58.7809105Z * [new branch] mlazos/buffers3 -> origin/mlazos/buffers3 2025-12-04T08:53:58.7809170Z * [new branch] mlazos/bwd -> origin/mlazos/bwd 2025-12-04T08:53:58.7809241Z * [new branch] mlazos/combo-test -> origin/mlazos/combo-test 2025-12-04T08:53:58.7809315Z * [new branch] mlazos/ctx-cleanup -> origin/mlazos/ctx-cleanup 2025-12-04T08:53:58.7809390Z * [new branch] mlazos/cuda-cmd-log -> origin/mlazos/cuda-cmd-log 2025-12-04T08:53:58.7809475Z * [new branch] mlazos/cudagraph-tests -> origin/mlazos/cudagraph-tests 2025-12-04T08:53:58.7809579Z * [new branch] mlazos/cudagraphs-measurement -> origin/mlazos/cudagraphs-measurement 2025-12-04T08:53:58.7809652Z * [new branch] mlazos/cutlass-test -> origin/mlazos/cutlass-test 2025-12-04T08:53:58.7809734Z * [new branch] mlazos/cutlass-topo-bug -> origin/mlazos/cutlass-topo-bug 2025-12-04T08:53:58.7809856Z * [new branch] mlazos/dataclass-proxy -> origin/mlazos/dataclass-proxy 2025-12-04T08:53:58.7809924Z * [new branch] mlazos/dc-attrs -> origin/mlazos/dc-attrs 2025-12-04T08:53:58.7809994Z * [new branch] mlazos/dc-helion -> origin/mlazos/dc-helion 2025-12-04T08:53:58.7810061Z * [new branch] mlazos/dict-fix -> origin/mlazos/dict-fix 2025-12-04T08:53:58.7810132Z * [new branch] mlazos/disable-tf -> origin/mlazos/disable-tf 2025-12-04T08:53:58.7810201Z * [new branch] mlazos/dupe-fix -> origin/mlazos/dupe-fix 2025-12-04T08:53:58.7810273Z * [new branch] mlazos/dyn-batch -> origin/mlazos/dyn-batch 2025-12-04T08:53:58.7810337Z * [new branch] mlazos/evt -> origin/mlazos/evt 2025-12-04T08:53:58.7810419Z * [new branch] mlazos/extract-examples -> origin/mlazos/extract-examples 2025-12-04T08:53:58.7810490Z * [new branch] mlazos/foreach-op -> origin/mlazos/foreach-op 2025-12-04T08:53:58.7810554Z * [new branch] mlazos/fp8 -> origin/mlazos/fp8 2025-12-04T08:53:58.7810624Z * [new branch] mlazos/fp8-bias -> origin/mlazos/fp8-bias 2025-12-04T08:53:58.7810704Z * [new branch] mlazos/fp8-bias-fusion -> origin/mlazos/fp8-bias-fusion 2025-12-04T08:53:58.7810774Z * [new branch] mlazos/fp8-fixes -> origin/mlazos/fp8-fixes 2025-12-04T08:53:58.7810844Z * [new branch] mlazos/freezing -> origin/mlazos/freezing 2025-12-04T08:53:58.7810912Z * [new branch] mlazos/h-comp -> origin/mlazos/h-comp 2025-12-04T08:53:58.7810978Z * [new branch] mlazos/h-comp2 -> origin/mlazos/h-comp2 2025-12-04T08:53:58.7811045Z * [new branch] mlazos/hash-hop -> origin/mlazos/hash-hop 2025-12-04T08:53:58.7811106Z * [new branch] mlazos/hc -> origin/mlazos/hc 2025-12-04T08:53:58.7811178Z * [new branch] mlazos/hc-cycles -> origin/mlazos/hc-cycles 2025-12-04T08:53:58.7811245Z * [new branch] mlazos/hc-fixes -> origin/mlazos/hc-fixes 2025-12-04T08:53:58.7811313Z * [new branch] mlazos/hc-fixes3 -> origin/mlazos/hc-fixes3 2025-12-04T08:53:58.7811383Z * [new branch] mlazos/hc-fixes4 -> origin/mlazos/hc-fixes4 2025-12-04T08:53:58.7811449Z * [new branch] mlazos/hc-hf -> origin/mlazos/hc-hf 2025-12-04T08:53:58.7811515Z * [new branch] mlazos/hc-mut -> origin/mlazos/hc-mut 2025-12-04T08:53:58.7811610Z * [new branch] mlazos/hc10 -> origin/mlazos/hc10 2025-12-04T08:53:58.7811673Z * [new branch] mlazos/hc11 -> origin/mlazos/hc11 2025-12-04T08:53:58.7811735Z * [new branch] mlazos/hc12 -> origin/mlazos/hc12 2025-12-04T08:53:58.7811800Z * [new branch] mlazos/hc13 -> origin/mlazos/hc13 2025-12-04T08:53:58.7811860Z * [new branch] mlazos/hc14 -> origin/mlazos/hc14 2025-12-04T08:53:58.7811920Z * [new branch] mlazos/hc15 -> origin/mlazos/hc15 2025-12-04T08:53:58.7811982Z * [new branch] mlazos/hc2 -> origin/mlazos/hc2 2025-12-04T08:53:58.7812044Z * [new branch] mlazos/hc4 -> origin/mlazos/hc4 2025-12-04T08:53:58.7812105Z * [new branch] mlazos/hc5 -> origin/mlazos/hc5 2025-12-04T08:53:58.7812168Z * [new branch] mlazos/hc6 -> origin/mlazos/hc6 2025-12-04T08:53:58.7812229Z * [new branch] mlazos/hc7 -> origin/mlazos/hc7 2025-12-04T08:53:58.7812290Z * [new branch] mlazos/hc8 -> origin/mlazos/hc8 2025-12-04T08:53:58.7812350Z * [new branch] mlazos/hc9 -> origin/mlazos/hc9 2025-12-04T08:53:58.7812448Z * [new branch] mlazos/hc_baseline2 -> origin/mlazos/hc_baseline2 2025-12-04T08:53:58.7812530Z * [new branch] mlazos/inductor-streams -> origin/mlazos/inductor-streams 2025-12-04T08:53:58.7812594Z * [new branch] mlazos/main -> origin/mlazos/main 2025-12-04T08:53:58.7812655Z * [new branch] mlazos/mcg2 -> origin/mlazos/mcg2 2025-12-04T08:53:58.7812728Z * [new branch] mlazos/meta-guards -> origin/mlazos/meta-guards 2025-12-04T08:53:58.7812833Z * [new branch] mlazos/mlazos/foreach-map-adam -> origin/mlazos/mlazos/foreach-map-adam 2025-12-04T08:53:58.7812932Z * [new branch] mlazos/mlazos/tf-mode-backup -> origin/mlazos/mlazos/tf-mode-backup 2025-12-04T08:53:58.7813000Z * [new branch] mlazos/mod-fix -> origin/mlazos/mod-fix 2025-12-04T08:53:58.7813067Z * [new branch] mlazos/mode-fix -> origin/mlazos/mode-fix 2025-12-04T08:53:58.7813135Z * [new branch] mlazos/offsets -> origin/mlazos/offsets 2025-12-04T08:53:58.7813211Z * [new branch] mlazos/overguarding -> origin/mlazos/overguarding 2025-12-04T08:53:58.7813284Z * [new branch] mlazos/proxy-ctors -> origin/mlazos/proxy-ctors 2025-12-04T08:53:58.7813353Z * [new branch] mlazos/quant-fix -> origin/mlazos/quant-fix 2025-12-04T08:53:58.7813423Z * [new branch] mlazos/resnet-fix -> origin/mlazos/resnet-fix 2025-12-04T08:53:58.7813497Z * [new branch] mlazos/rm-buf-names -> origin/mlazos/rm-buf-names 2025-12-04T08:53:58.7813566Z * [new branch] mlazos/rm-code -> origin/mlazos/rm-code 2025-12-04T08:53:58.7813634Z * [new branch] mlazos/rm-spam -> origin/mlazos/rm-spam 2025-12-04T08:53:58.7813698Z * [new branch] mlazos/rtp -> origin/mlazos/rtp 2025-12-04T08:53:58.7813778Z * [new branch] mlazos/static-idx-dbg -> origin/mlazos/static-idx-dbg 2025-12-04T08:53:58.7813864Z * [new branch] mlazos/static-inputs-log -> origin/mlazos/static-inputs-log 2025-12-04T08:53:58.7813931Z * [new branch] mlazos/stests -> origin/mlazos/stests 2025-12-04T08:53:58.7814003Z * [new branch] mlazos/stream-ops -> origin/mlazos/stream-ops 2025-12-04T08:53:58.7814068Z * [new branch] mlazos/td-fix2 -> origin/mlazos/td-fix2 2025-12-04T08:53:58.7814146Z * [new branch] mlazos/tensor-hasattr2 -> origin/mlazos/tensor-hasattr2 2025-12-04T08:53:58.7814236Z * [new branch] mlazos/test -> origin/mlazos/test 2025-12-04T08:53:58.7814304Z * [new branch] mlazos/tf-mode -> origin/mlazos/tf-mode 2025-12-04T08:53:58.7814382Z * [new branch] mlazos/tf-mode-backup2 -> origin/mlazos/tf-mode-backup2 2025-12-04T08:53:58.7814460Z * [new branch] mlazos/tf-mode-reland -> origin/mlazos/tf-mode-reland 2025-12-04T08:53:58.7814537Z * [new branch] mlazos/tf-mode-reland2 -> origin/mlazos/tf-mode-reland2 2025-12-04T08:53:58.7814613Z * [new branch] mlazos/tf-mode-reland3 -> origin/mlazos/tf-mode-reland3 2025-12-04T08:53:58.7814690Z * [new branch] mlazos/triton-no-epi -> origin/mlazos/triton-no-epi 2025-12-04T08:53:58.7814762Z * [new branch] mlazos/tune-proto -> origin/mlazos/tune-proto 2025-12-04T08:53:58.7814834Z * [new branch] mlazos/tuple-fixes -> origin/mlazos/tuple-fixes 2025-12-04T08:53:58.7814913Z * [new branch] mlazos/tuple-fixes2 -> origin/mlazos/tuple-fixes2 2025-12-04T08:53:58.7814990Z * [new branch] mlazos/tuple-handling -> origin/mlazos/tuple-handling 2025-12-04T08:53:58.7815072Z * [new branch] mlazos/user-stream-base -> origin/mlazos/user-stream-base 2025-12-04T08:53:58.7815177Z * [new branch] mlazos/user-streams -> origin/mlazos/user-streams 2025-12-04T08:53:58.7815269Z * [new branch] mlazos/user-streams-backup -> origin/mlazos/user-streams-backup 2025-12-04T08:53:58.7815364Z * [new branch] mlazos/user-streams-backup2 -> origin/mlazos/user-streams-backup2 2025-12-04T08:53:58.7815436Z * [new branch] mlazos/vary-beta -> origin/mlazos/vary-beta 2025-12-04T08:53:58.7815507Z * [new branch] mlazos/vary-beta2 -> origin/mlazos/vary-beta2 2025-12-04T08:53:58.7815580Z * [new branch] mlazos/weird-perf1 -> origin/mlazos/weird-perf1 2025-12-04T08:53:58.7815655Z * [new branch] mm_out_dtype_compile -> origin/mm_out_dtype_compile 2025-12-04T08:53:58.7815721Z * [new branch] module-shim -> origin/module-shim 2025-12-04T08:53:58.7815783Z * [new branch] move_config -> origin/move_config 2025-12-04T08:53:58.7815855Z * [new branch] msaroufim/reduce -> origin/msaroufim/reduce 2025-12-04T08:53:58.7815924Z * [new branch] mtia/basic-cmake -> origin/mtia/basic-cmake 2025-12-04T08:53:58.7816025Z * [new branch] mwizak/fix-triton-block-shape -> origin/mwizak/fix-triton-block-shape 2025-12-04T08:53:58.7816093Z * [new branch] my_varlen_backup -> origin/my_varlen_backup 2025-12-04T08:53:58.7816167Z * [new branch] nativert_num_outputs -> origin/nativert_num_outputs 2025-12-04T08:53:58.7816231Z * [new branch] new-codegen -> origin/new-codegen 2025-12-04T08:53:58.7816301Z * [new branch] newtest-base -> origin/newtest-base 2025-12-04T08:53:58.7816373Z * [new branch] ngimel/addmm_dtype -> origin/ngimel/addmm_dtype 2025-12-04T08:53:58.7816440Z * [new branch] ngimel/div_inv -> origin/ngimel/div_inv 2025-12-04T08:53:58.7816521Z * [new branch] ngimel/error_index_list -> origin/ngimel/error_index_list 2025-12-04T08:53:58.7816593Z * [new branch] ngimel/gather_grid -> origin/ngimel/gather_grid 2025-12-04T08:53:58.7816682Z * [new branch] ngimel/gather_grid_release -> origin/ngimel/gather_grid_release 2025-12-04T08:53:58.7816798Z * [new branch] ngimel/gg_new -> origin/ngimel/gg_new 2025-12-04T08:53:58.7816868Z * [new branch] ngimel/hostalloc -> origin/ngimel/hostalloc 2025-12-04T08:53:58.7816941Z * [new branch] ngimel/storage_id -> origin/ngimel/storage_id 2025-12-04T08:53:58.7817040Z * [new branch] nightly -> origin/nightly 2025-12-04T08:53:58.7817157Z * [new branch] nikitaved/addmm_1_rowcol_lt_path_check -> origin/nikitaved/addmm_1_rowcol_lt_path_check 2025-12-04T08:53:58.7817283Z * [new branch] nikitaved/addmm_epilogue_fusions_2d_bias -> origin/nikitaved/addmm_epilogue_fusions_2d_bias 2025-12-04T08:53:58.7817413Z * [new branch] nikitaved/addmm_epilogue_fusions_inductor -> origin/nikitaved/addmm_epilogue_fusions_inductor 2025-12-04T08:53:58.7817536Z * [new branch] nikitaved/addmm_epilogue_fusions_scratch -> origin/nikitaved/addmm_epilogue_fusions_scratch 2025-12-04T08:53:58.7817654Z * [new branch] nikitaved/grad_addmm_epilogue_fusions -> origin/nikitaved/grad_addmm_epilogue_fusions 2025-12-04T08:53:58.7817768Z * [new branch] nikitaved/simpler_can_use_32bit_index -> origin/nikitaved/simpler_can_use_32bit_index 2025-12-04T08:53:58.7817837Z * [new branch] nikitaved/test -> origin/nikitaved/test 2025-12-04T08:53:58.7817968Z * [new branch] nmacchioni-perf-test-async-autotune -> origin/nmacchioni-perf-test-async-autotune 2025-12-04T08:53:58.7818047Z * [new branch] no_distributed_log_spew -> origin/no_distributed_log_spew 2025-12-04T08:53:58.7818158Z * [new branch] nofun-hack -> origin/nofun-hack 2025-12-04T08:53:58.7818221Z * [new branch] norm_bench -> origin/norm_bench 2025-12-04T08:53:58.7818296Z * [new branch] nullplay/fuse_matmul -> origin/nullplay/fuse_matmul 2025-12-04T08:53:58.7818371Z * [new branch] nullplay_fuse_matmul -> origin/nullplay_fuse_matmul 2025-12-04T08:53:58.7818439Z * [new branch] optimizer_test -> origin/optimizer_test 2025-12-04T08:53:58.7818508Z * [new branch] orig/release/1.10 -> origin/orig/release/1.10 2025-12-04T08:53:58.7818579Z * [new branch] orig/release/1.11 -> origin/orig/release/1.11 2025-12-04T08:53:58.7818648Z * [new branch] orig/release/1.12 -> origin/orig/release/1.12 2025-12-04T08:53:58.7818716Z * [new branch] orig/release/1.13 -> origin/orig/release/1.13 2025-12-04T08:53:58.7818784Z * [new branch] orig/release/1.6 -> origin/orig/release/1.6 2025-12-04T08:53:58.7818851Z * [new branch] orig/release/1.7 -> origin/orig/release/1.7 2025-12-04T08:53:58.7818917Z * [new branch] orig/release/1.8 -> origin/orig/release/1.8 2025-12-04T08:53:58.7818984Z * [new branch] orig/release/1.9 -> origin/orig/release/1.9 2025-12-04T08:53:58.7819050Z * [new branch] orig/release/2.0 -> origin/orig/release/2.0 2025-12-04T08:53:58.7819114Z * [new branch] orig/release/2.1 -> origin/orig/release/2.1 2025-12-04T08:53:58.7819182Z * [new branch] orig/release/2.2 -> origin/orig/release/2.2 2025-12-04T08:53:58.7819247Z * [new branch] orig/release/2.3 -> origin/orig/release/2.3 2025-12-04T08:53:58.7819311Z * [new branch] orig/release/2.4 -> origin/orig/release/2.4 2025-12-04T08:53:58.7819380Z * [new branch] orig/release/2.5 -> origin/orig/release/2.5 2025-12-04T08:53:58.7819447Z * [new branch] orig/release/2.6 -> origin/orig/release/2.6 2025-12-04T08:53:58.7819512Z * [new branch] orig/release/2.7 -> origin/orig/release/2.7 2025-12-04T08:53:58.7819578Z * [new branch] orig/release/2.8 -> origin/orig/release/2.8 2025-12-04T08:53:58.7819643Z * [new branch] orig/release/2.9 -> origin/orig/release/2.9 2025-12-04T08:53:58.7819728Z * [new branch] origin/gh/fxdawnn/1/base -> origin/origin/gh/fxdawnn/1/base 2025-12-04T08:53:58.7819845Z * [new branch] origin/gh/fxdawnn/1/orig -> origin/origin/gh/fxdawnn/1/orig 2025-12-04T08:53:58.7819928Z * [new branch] origin/gh/zpcore/14/orig -> origin/origin/gh/zpcore/14/orig 2025-12-04T08:53:58.7820000Z * [new branch] oulgen-patch-1 -> origin/oulgen-patch-1 2025-12-04T08:53:58.7820069Z * [new branch] oulgen-patch-2 -> origin/oulgen-patch-2 2025-12-04T08:53:58.7820139Z * [new branch] oulgen-patch-3 -> origin/oulgen-patch-3 2025-12-04T08:53:58.7820206Z * [new branch] oulgen-patch-4 -> origin/oulgen-patch-4 2025-12-04T08:53:58.7820276Z * [new branch] padded-tensor -> origin/padded-tensor 2025-12-04T08:53:58.7820339Z * [new branch] pca2 -> origin/pca2 2025-12-04T08:53:58.7820416Z * [new branch] per_channel_backup -> origin/per_channel_backup 2025-12-04T08:53:58.7820480Z * [new branch] perf_ops -> origin/perf_ops 2025-12-04T08:53:58.7820548Z * [new branch] perf_ops_2_9 -> origin/perf_ops_2_9 2025-12-04T08:53:58.7820625Z * [new branch] pianpwk-patch-1 -> origin/pianpwk-patch-1 2025-12-04T08:53:58.7820712Z * [new branch] pianpwk/__draft_debug_mode -> origin/pianpwk/__draft_debug_mode 2025-12-04T08:53:58.7820851Z * [new branch] pianpwk/_debug_mode_for_triton_draft -> origin/pianpwk/_debug_mode_for_triton_draft 2025-12-04T08:53:58.7820957Z * [new branch] pianpwk/_debug_nn_module_compile -> origin/pianpwk/_debug_nn_module_compile 2025-12-04T08:53:58.7821046Z * [new branch] pianpwk/_draft_triton_11_3 -> origin/pianpwk/_draft_triton_11_3 2025-12-04T08:53:58.7821140Z * [new branch] pianpwk/_manual_bucket_draft -> origin/pianpwk/_manual_bucket_draft 2025-12-04T08:53:58.7821245Z * [new branch] pianpwk/_profile_w_dispatch_keys -> origin/pianpwk/_profile_w_dispatch_keys 2025-12-04T08:53:58.7821345Z * [new branch] pianpwk/_super_draft_debug_mode -> origin/pianpwk/_super_draft_debug_mode 2025-12-04T08:53:58.7821452Z * [new branch] pianpwk/_unbacked_local_shard_size -> origin/pianpwk/_unbacked_local_shard_size 2025-12-04T08:53:58.7821529Z * [new branch] pianpwk/anomaly_tb -> origin/pianpwk/anomaly_tb 2025-12-04T08:53:58.7821611Z * [new branch] pianpwk/auto_fx_annotate -> origin/pianpwk/auto_fx_annotate 2025-12-04T08:53:58.7821726Z * [new branch] pianpwk/backed_size_oblivious_export -> origin/pianpwk/backed_size_oblivious_export 2025-12-04T08:53:58.7821813Z * [new branch] pianpwk/bert_dynamic_perf -> origin/pianpwk/bert_dynamic_perf 2025-12-04T08:53:58.7821912Z * [new branch] pianpwk/debug_fwd_stack_traces -> origin/pianpwk/debug_fwd_stack_traces 2025-12-04T08:53:58.7821998Z * [new branch] pianpwk/debug_hash_tensor -> origin/pianpwk/debug_hash_tensor 2025-12-04T08:53:58.7822091Z * [new branch] pianpwk/debug_mode_annotate -> origin/pianpwk/debug_mode_annotate 2025-12-04T08:53:58.7822182Z * [new branch] pianpwk/debug_mode_defaults -> origin/pianpwk/debug_mode_defaults 2025-12-04T08:53:58.7822263Z * [new branch] pianpwk/debug_mode_hacks -> origin/pianpwk/debug_mode_hacks 2025-12-04T08:53:58.7822371Z * [new branch] pianpwk/debug_mode_opcall_refactor -> origin/pianpwk/debug_mode_opcall_refactor 2025-12-04T08:53:58.7822460Z * [new branch] pianpwk/debug_mode_show_ids -> origin/pianpwk/debug_mode_show_ids 2025-12-04T08:53:58.7822544Z * [new branch] pianpwk/debug_mode_triton -> origin/pianpwk/debug_mode_triton 2025-12-04T08:53:58.7822640Z * [new branch] pianpwk/debug_show_stack_trace -> origin/pianpwk/debug_show_stack_trace 2025-12-04T08:53:58.7822742Z * [new branch] pianpwk/debug_wait_on_collective -> origin/pianpwk/debug_wait_on_collective 2025-12-04T08:53:58.7822866Z * [new branch] pianpwk/debugmode_compile_tf -> origin/pianpwk/debugmode_compile_tf 2025-12-04T08:53:58.7822992Z * [new branch] pianpwk/dispatch_key_debugging_for_debug -> origin/pianpwk/dispatch_key_debugging_for_debug 2025-12-04T08:53:58.7823101Z * [new branch] pianpwk/draft_debug_mode_tfcompile -> origin/pianpwk/draft_debug_mode_tfcompile 2025-12-04T08:53:58.7823196Z * [new branch] pianpwk/draft_multikernel_nn -> origin/pianpwk/draft_multikernel_nn 2025-12-04T08:53:58.7823312Z * [new branch] pianpwk/draft_multikernel_status_10_5 -> origin/pianpwk/draft_multikernel_status_10_5 2025-12-04T08:53:58.7823404Z * [new branch] pianpwk/dtensor_custom_chunk -> origin/pianpwk/dtensor_custom_chunk 2025-12-04T08:53:58.7823508Z * [new branch] pianpwk/dtensor_unbacked_keypath -> origin/pianpwk/dtensor_unbacked_keypath 2025-12-04T08:53:58.7823591Z * [new branch] pianpwk/event_list_tree -> origin/pianpwk/event_list_tree 2025-12-04T08:53:58.7823673Z * [new branch] pianpwk/false_numel_refs -> origin/pianpwk/false_numel_refs 2025-12-04T08:53:58.7823751Z * [new branch] pianpwk/maybe_guard_rel -> origin/pianpwk/maybe_guard_rel 2025-12-04T08:53:58.7823882Z * [new branch] pianpwk/multikernel_hints_draft -> origin/pianpwk/multikernel_hints_draft 2025-12-04T08:53:58.7823993Z * [new branch] pianpwk/no_size_oblivious_slice_scat -> origin/pianpwk/no_size_oblivious_slice_scat 2025-12-04T08:53:58.7824109Z * [new branch] pianpwk/oblivious_reshape_view_better -> origin/pianpwk/oblivious_reshape_view_better 2025-12-04T08:53:58.7824194Z * [new branch] pianpwk/pre_forward_hook -> origin/pianpwk/pre_forward_hook 2025-12-04T08:53:58.7824301Z * [new branch] pianpwk/skip_python_keys_alternate -> origin/pianpwk/skip_python_keys_alternate 2025-12-04T08:53:58.7824408Z * [new branch] pianpwk/skip_python_keys_in_guards -> origin/pianpwk/skip_python_keys_in_guards 2025-12-04T08:53:58.7824490Z * [new branch] pianpwk/sym_tokens_draft -> origin/pianpwk/sym_tokens_draft 2025-12-04T08:53:58.7824570Z * [new branch] pianpwk/symint_one_hot -> origin/pianpwk/symint_one_hot 2025-12-04T08:53:58.7824686Z * [new branch] pianpwk/test_pointwise_guard_or_false -> origin/pianpwk/test_pointwise_guard_or_false 2025-12-04T08:53:58.7824786Z * [new branch] pianpwk/totally_draft_sym_wrap -> origin/pianpwk/totally_draft_sym_wrap 2025-12-04T08:53:58.7824869Z * [new branch] pianpwk/try_dumb_stuff -> origin/pianpwk/try_dumb_stuff 2025-12-04T08:53:58.7824950Z * [new branch] pianpwk/try_dumb_stuff_2 -> origin/pianpwk/try_dumb_stuff_2 2025-12-04T08:53:58.7825042Z * [new branch] pianpwk/unbacked_dtensor_mm -> origin/pianpwk/unbacked_dtensor_mm 2025-12-04T08:53:58.7825141Z * [new branch] pianpwk/unbacked_tracing_12_2 -> origin/pianpwk/unbacked_tracing_12_2 2025-12-04T08:53:58.7825218Z * [new branch] pianpwk/user_symints -> origin/pianpwk/user_symints 2025-12-04T08:53:58.7825296Z * [new branch] pianpwk/wan21_reshape -> origin/pianpwk/wan21_reshape 2025-12-04T08:53:58.7825391Z * [new branch] piz/fix_partial_backward_1112 -> origin/piz/fix_partial_backward_1112 2025-12-04T08:53:58.7825469Z * [new branch] piz/prop_cache_clean -> origin/piz/prop_cache_clean 2025-12-04T08:53:58.7825537Z * [new branch] pool-separate -> origin/pool-separate 2025-12-04T08:53:58.7825599Z * [new branch] pr-156087 -> origin/pr-156087 2025-12-04T08:53:58.7825663Z * [new branch] pr/131860 -> origin/pr/131860 2025-12-04T08:53:58.7825732Z * [new branch] predispatch_to -> origin/predispatch_to 2025-12-04T08:53:58.7825826Z * [new branch] protect-c17 -> origin/protect-c17 2025-12-04T08:53:58.7825896Z * [new branch] pt-opt-cuda3 -> origin/pt-opt-cuda3 2025-12-04T08:53:58.7825977Z * [new branch] python_compiled_autograd -> origin/python_compiled_autograd 2025-12-04T08:53:58.7826108Z * [new branch] q1l1/fix_device_moved_constant_type_unknown -> origin/q1l1/fix_device_moved_constant_type_unknown 2025-12-04T08:53:58.7826248Z * [new branch] q1l1/fix_wrong_default_type_for_kernel_call_args -> origin/q1l1/fix_wrong_default_type_for_kernel_call_args 2025-12-04T08:53:58.7826327Z * [new branch] qchip/export-D54134695 -> origin/qchip/export-D54134695 2025-12-04T08:53:58.7826402Z * [new branch] quote-pytest_cache -> origin/quote-pytest_cache 2025-12-04T08:53:58.7826500Z * [new branch] reland-accgrad-stream-warn -> origin/reland-accgrad-stream-warn 2025-12-04T08:53:58.7826567Z * [new branch] release/1.10 -> origin/release/1.10 2025-12-04T08:53:58.7826633Z * [new branch] release/1.11 -> origin/release/1.11 2025-12-04T08:53:58.7826696Z * [new branch] release/1.12 -> origin/release/1.12 2025-12-04T08:53:58.7826806Z * [new branch] release/1.13 -> origin/release/1.13 2025-12-04T08:53:58.7826909Z * [new branch] release/1.4 -> origin/release/1.4 2025-12-04T08:53:58.7826974Z * [new branch] release/1.4.1 -> origin/release/1.4.1 2025-12-04T08:53:58.7827037Z * [new branch] release/1.5 -> origin/release/1.5 2025-12-04T08:53:58.7827100Z * [new branch] release/1.6 -> origin/release/1.6 2025-12-04T08:53:58.7827161Z * [new branch] release/1.7 -> origin/release/1.7 2025-12-04T08:53:58.7827223Z * [new branch] release/1.8 -> origin/release/1.8 2025-12-04T08:53:58.7827289Z * [new branch] release/1.9 -> origin/release/1.9 2025-12-04T08:53:58.7827350Z * [new branch] release/2.0 -> origin/release/2.0 2025-12-04T08:53:58.7827410Z * [new branch] release/2.1 -> origin/release/2.1 2025-12-04T08:53:58.7827470Z * [new branch] release/2.2 -> origin/release/2.2 2025-12-04T08:53:58.7827531Z * [new branch] release/2.3 -> origin/release/2.3 2025-12-04T08:53:58.7827592Z * [new branch] release/2.4 -> origin/release/2.4 2025-12-04T08:53:58.7827654Z * [new branch] release/2.5 -> origin/release/2.5 2025-12-04T08:53:58.7827715Z * [new branch] release/2.6 -> origin/release/2.6 2025-12-04T08:53:58.7827776Z * [new branch] release/2.7 -> origin/release/2.7 2025-12-04T08:53:58.7827836Z * [new branch] release/2.8 -> origin/release/2.8 2025-12-04T08:53:58.7827899Z * [new branch] release/2.9 -> origin/release/2.9 2025-12-04T08:53:58.7827964Z * [new branch] release_notes -> origin/release_notes 2025-12-04T08:53:58.7828041Z * [new branch] remove_pyinterpreter -> origin/remove_pyinterpreter 2025-12-04T08:53:58.7828166Z * [new branch] replace-pytorch-labs-20250812-195836 -> origin/replace-pytorch-labs-20250812-195836 2025-12-04T08:53:58.7828290Z * [new branch] replace-pytorch-labs-20250812-200248 -> origin/replace-pytorch-labs-20250812-200248 2025-12-04T08:53:58.7828409Z * [new branch] replace-pytorch-labs-20250812-200324 -> origin/replace-pytorch-labs-20250812-200324 2025-12-04T08:53:58.7828529Z * [new branch] replace-pytorch-labs-20250812-204020 -> origin/replace-pytorch-labs-20250812-204020 2025-12-04T08:53:58.7828664Z * [new branch] revert-131069-gh/krzysztofjordan/1/head -> origin/revert-131069-gh/krzysztofjordan/1/head 2025-12-04T08:53:58.7828812Z * [new branch] revert-131469-gh/andrewor14/51/head -> origin/revert-131469-gh/andrewor14/51/head 2025-12-04T08:53:58.7828915Z * [new branch] revert-152361-gh/fadara01/1/head -> origin/revert-152361-gh/fadara01/1/head 2025-12-04T08:53:58.7829020Z * [new branch] revert-156870-gh/skarjala/3/head -> origin/revert-156870-gh/skarjala/3/head 2025-12-04T08:53:58.7829191Z * [new branch] revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ -> origin/revert-157914-cherry-pick-157503-by-pytorch_bot_bot_ 2025-12-04T08:53:58.7829287Z * [new branch] revert-hoo-invoke-subgraph -> origin/revert-hoo-invoke-subgraph 2025-12-04T08:53:58.7829388Z * [new branch] revert_always_build_distributed -> origin/revert_always_build_distributed 2025-12-04T08:53:58.7829455Z * [new branch] rms_norm_patch -> origin/rms_norm_patch 2025-12-04T08:53:58.7829555Z * [new branch] ruisi/fix_all_to_all_estimation -> origin/ruisi/fix_all_to_all_estimation 2025-12-04T08:53:58.7829641Z * [new branch] ruisi/fix_comm_estimation -> origin/ruisi/fix_comm_estimation 2025-12-04T08:53:58.7829750Z * [new branch] ruisi/fix_dynamic_shape_estimation -> origin/ruisi/fix_dynamic_shape_estimation 2025-12-04T08:53:58.7829875Z * [new branch] ruisi/fix_llama3_autobucketing -> origin/ruisi/fix_llama3_autobucketing 2025-12-04T08:53:58.7829982Z * [new branch] ruisi/fix_manual_bucketing_ep_pass -> origin/ruisi/fix_manual_bucketing_ep_pass 2025-12-04T08:53:58.7830067Z * [new branch] ruisi/manual_bucket_pass -> origin/ruisi/manual_bucket_pass 2025-12-04T08:53:58.7830217Z * [new branch] ryanguo99/cleanup-dynamo-expected-failures -> origin/ryanguo99/cleanup-dynamo-expected-failures 2025-12-04T08:53:58.7830304Z * [new branch] ryanguo99/fix-closure-var -> origin/ryanguo99/fix-closure-var 2025-12-04T08:53:58.7830385Z * [new branch] rzou/faketensor_bench -> origin/rzou/faketensor_bench 2025-12-04T08:53:58.7830449Z * [new branch] rzou/njt -> origin/rzou/njt 2025-12-04T08:53:58.7830512Z * [new branch] rzou/pca -> origin/rzou/pca 2025-12-04T08:53:58.7830579Z * [new branch] rzou/realprop -> origin/rzou/realprop 2025-12-04T08:53:58.7830645Z * [new branch] samplevllm -> origin/samplevllm 2025-12-04T08:53:58.7830813Z * [new branch] sanchitintel/weird_thing_with_test_cpu_select_algorithm -> origin/sanchitintel/weird_thing_with_test_cpu_select_algorithm 2025-12-04T08:53:58.7830907Z * [new branch] sapling-pr-archive-SS-JIA -> origin/sapling-pr-archive-SS-JIA 2025-12-04T08:53:58.7831021Z * [new branch] sapling-pr-archive-tushar00jain -> origin/sapling-pr-archive-tushar00jain 2025-12-04T08:53:58.7831084Z * [new branch] save -> origin/save 2025-12-04T08:53:58.7831147Z * [new branch] scaled_mm -> origin/scaled_mm 2025-12-04T08:53:58.7831214Z * [new branch] scan_attempt -> origin/scan_attempt 2025-12-04T08:53:58.7831277Z * [new branch] sdym/2.5.1 -> origin/sdym/2.5.1 2025-12-04T08:53:58.7831389Z * [new branch] sekyondaMeta-dynamoconfig-fix -> origin/sekyondaMeta-dynamoconfig-fix 2025-12-04T08:53:58.7831465Z * [new branch] shengf/fx-xform-perf -> origin/shengf/fx-xform-perf 2025-12-04T08:53:58.7831543Z * [new branch] shoumikhin-patch-1 -> origin/shoumikhin-patch-1 2025-12-04T08:53:58.7831620Z * [new branch] solve-accuracy-fix -> origin/solve-accuracy-fix 2025-12-04T08:53:58.7831699Z * [new branch] some_rocm_inductor_skips -> origin/some_rocm_inductor_skips 2025-12-04T08:53:58.7831815Z * [new branch] soulitzer/stash-tls-ac -> origin/soulitzer/stash-tls-ac 2025-12-04T08:53:58.7831900Z * [new branch] sparse-mm-bf16-support -> origin/sparse-mm-bf16-support 2025-12-04T08:53:58.7831974Z * [new branch] starterTaskUpdate -> origin/starterTaskUpdate 2025-12-04T08:53:58.7832035Z * [new branch] suo -> origin/suo 2025-12-04T08:53:58.7832099Z * [new branch] sve-poc -> origin/sve-poc 2025-12-04T08:53:58.7832161Z * [new branch] switch-bn -> origin/switch-bn 2025-12-04T08:53:58.7832255Z * [new branch] sy_annotation_in_autograd_hop -> origin/sy_annotation_in_autograd_hop 2025-12-04T08:53:58.7832326Z * [new branch] sy_aot_eager_record -> origin/sy_aot_eager_record 2025-12-04T08:53:58.7832395Z * [new branch] sy_custom_bucketing -> origin/sy_custom_bucketing 2025-12-04T08:53:58.7832465Z * [new branch] sy_debug_mode_test -> origin/sy_debug_mode_test 2025-12-04T08:53:58.7832532Z * [new branch] sy_deserialize -> origin/sy_deserialize 2025-12-04T08:53:58.7832598Z * [new branch] sy_dump_gm_code -> origin/sy_dump_gm_code 2025-12-04T08:53:58.7832663Z * [new branch] sy_exp -> origin/sy_exp 2025-12-04T08:53:58.7832761Z * [new branch] sy_export_annotation -> origin/sy_export_annotation 2025-12-04T08:53:58.7832832Z * [new branch] sy_invoke_subgraph -> origin/sy_invoke_subgraph 2025-12-04T08:53:58.7832902Z * [new branch] sy_kernel_bw_name -> origin/sy_kernel_bw_name 2025-12-04T08:53:58.7832969Z * [new branch] sy_multi_arch -> origin/sy_multi_arch 2025-12-04T08:53:58.7833036Z * [new branch] sy_nn_module_stack -> origin/sy_nn_module_stack 2025-12-04T08:53:58.7833109Z * [new branch] sy_original_dtensor -> origin/sy_original_dtensor 2025-12-04T08:53:58.7833178Z * [new branch] sy_profiler_cia -> origin/sy_profiler_cia 2025-12-04T08:53:58.7833243Z * [new branch] symm_mem_sync -> origin/symm_mem_sync 2025-12-04T08:53:58.7833329Z * [new branch] sympy-bottleneck-repro -> origin/sympy-bottleneck-repro 2025-12-04T08:53:58.7833410Z * [new branch] tensordict_integration -> origin/tensordict_integration 2025-12-04T08:53:58.7833489Z * [new branch] test-move-conda-builds -> origin/test-move-conda-builds 2025-12-04T08:53:58.7833552Z * [new branch] test-old -> origin/test-old 2025-12-04T08:53:58.7833617Z * [new branch] test/bmm_heur -> origin/test/bmm_heur 2025-12-04T08:53:58.7833715Z * [new branch] tianren/customOp_autotune_fix -> origin/tianren/customOp_autotune_fix 2025-12-04T08:53:58.7833830Z * [new branch] tianren/customOp_enable_max_autotune -> origin/tianren/customOp_enable_max_autotune 2025-12-04T08:53:58.7833912Z * [new branch] tianren/customOp_fusion -> origin/tianren/customOp_fusion 2025-12-04T08:53:58.7834038Z * [new branch] tianren/customop_collectiveop_benchmark -> origin/tianren/customop_collectiveop_benchmark 2025-12-04T08:53:58.7834177Z * [new branch] tianren/customop_collectiveop_benchmark_fix -> origin/tianren/customop_collectiveop_benchmark_fix 2025-12-04T08:53:58.7834280Z * [new branch] tianren/customop_dynamic_config -> origin/tianren/customop_dynamic_config 2025-12-04T08:53:58.7834371Z * [new branch] tianren/dynamic_range_input -> origin/tianren/dynamic_range_input 2025-12-04T08:53:58.7834471Z * [new branch] tianren/dynamic_range_input_fix -> origin/tianren/dynamic_range_input_fix 2025-12-04T08:53:58.7834575Z * [new branch] tianren/dynamic_range_input_merge -> origin/tianren/dynamic_range_input_merge 2025-12-04T08:53:58.7834701Z * [new branch] tianren/flex_paged_attn_fix_temp -> origin/tianren/flex_paged_attn_fix_temp 2025-12-04T08:53:58.7834780Z * [new branch] tianren/fx_codegen_dump -> origin/tianren/fx_codegen_dump 2025-12-04T08:53:58.7834864Z * [new branch] tianren/symmetric_memory -> origin/tianren/symmetric_memory 2025-12-04T08:53:58.7834931Z * [new branch] tianren/test -> origin/tianren/test 2025-12-04T08:53:58.7835006Z * [new branch] tidy_performance_cyy -> origin/tidy_performance_cyy 2025-12-04T08:53:58.7835065Z * [new branch] tmp -> origin/tmp 2025-12-04T08:53:58.7835133Z * [new branch] torchtitan_ep -> origin/torchtitan_ep 2025-12-04T08:53:58.7835210Z * [new branch] torchtitan_integration -> origin/torchtitan_integration 2025-12-04T08:53:58.7835293Z * [new branch] trace_fsdp_torchtune_lora -> origin/trace_fsdp_torchtune_lora 2025-12-04T08:53:58.7835379Z * [new branch] traceable_fsdp_unit_tests -> origin/traceable_fsdp_unit_tests 2025-12-04T08:53:58.7835450Z * [new branch] tree_loop_vec_base -> origin/tree_loop_vec_base 2025-12-04T08:53:58.7835515Z * [new branch] triton_kernel -> origin/triton_kernel 2025-12-04T08:53:58.7835606Z * [new branch] tt_pkg_1908 -> origin/tt_pkg_1908 2025-12-04T08:53:58.7835667Z * [new branch] type_dec -> origin/type_dec 2025-12-04T08:53:58.7835760Z * [new branch] udate-sphinx-dependancies -> origin/udate-sphinx-dependancies 2025-12-04T08:53:58.7835898Z * [new branch] update-audio-commit-hash/17630256502-1803-1 -> origin/update-audio-commit-hash/17630256502-1803-1 2025-12-04T08:53:58.7836031Z * [new branch] update-audio-commit-hash/19087141161-1916-1 -> origin/update-audio-commit-hash/19087141161-1916-1 2025-12-04T08:53:58.7836163Z * [new branch] update-audio-commit-hash/19250643381-1929-1 -> origin/update-audio-commit-hash/19250643381-1929-1 2025-12-04T08:53:58.7836294Z * [new branch] update-audio-commit-hash/19397724337-1935-1 -> origin/update-audio-commit-hash/19397724337-1935-1 2025-12-04T08:53:58.7836423Z * [new branch] update-audio-commit-hash/19555670148-1941-1 -> origin/update-audio-commit-hash/19555670148-1941-1 2025-12-04T08:53:58.7836553Z * [new branch] update-audio-commit-hash/19750627930-1946-1 -> origin/update-audio-commit-hash/19750627930-1946-1 2025-12-04T08:53:58.7836686Z * [new branch] update-triton-commit-hash/13663274526-1487-2 -> origin/update-triton-commit-hash/13663274526-1487-2 2025-12-04T08:53:58.7836867Z * [new branch] update-vision-commit-hash/19087141161-1916-1 -> origin/update-vision-commit-hash/19087141161-1916-1 2025-12-04T08:53:58.7837003Z * [new branch] update-vision-commit-hash/19184897099-1925-1 -> origin/update-vision-commit-hash/19184897099-1925-1 2025-12-04T08:53:58.7837137Z * [new branch] update-vision-commit-hash/19250643381-1929-1 -> origin/update-vision-commit-hash/19250643381-1929-1 2025-12-04T08:53:58.7837269Z * [new branch] update-vision-commit-hash/19381328640-1934-1 -> origin/update-vision-commit-hash/19381328640-1934-1 2025-12-04T08:53:58.7837404Z * [new branch] update-vision-commit-hash/19485237164-1938-1 -> origin/update-vision-commit-hash/19485237164-1938-1 2025-12-04T08:53:58.7837533Z * [new branch] update-vllm-commit-hash/18451675449-1879-1 -> origin/update-vllm-commit-hash/18451675449-1879-1 2025-12-04T08:53:58.7837619Z * [new branch] update-vllm-dockerfile -> origin/update-vllm-dockerfile 2025-12-04T08:53:58.7837742Z * [new branch] update-xla-commit-hash/19224287370-211-1 -> origin/update-xla-commit-hash/19224287370-211-1 2025-12-04T08:53:58.7837905Z * [new branch] update-xla-commit-hash/19422028566-212-1 -> origin/update-xla-commit-hash/19422028566-212-1 2025-12-04T08:53:58.7838028Z * [new branch] update-xla-commit-hash/19626841311-213-1 -> origin/update-xla-commit-hash/19626841311-213-1 2025-12-04T08:53:58.7838154Z * [new branch] update_docs_torch_multinomial_issue#125388 -> origin/update_docs_torch_multinomial_issue#125388 2025-12-04T08:53:58.7838233Z * [new branch] update_operator_readme -> origin/update_operator_readme 2025-12-04T08:53:58.7838324Z * [new branch] update_slow_tests_1722488736 -> origin/update_slow_tests_1722488736 2025-12-04T08:53:58.7838413Z * [new branch] update_slow_tests_1722879173 -> origin/update_slow_tests_1722879173 2025-12-04T08:53:58.7838500Z * [new branch] update_slow_tests_1762155677 -> origin/update_slow_tests_1762155677 2025-12-04T08:53:58.7838587Z * [new branch] update_slow_tests_1763365283 -> origin/update_slow_tests_1763365283 2025-12-04T08:53:58.7838668Z * [new branch] update_submodule_FBGEMM -> origin/update_submodule_FBGEMM 2025-12-04T08:53:58.7838746Z * [new branch] update_submodule_kineto -> origin/update_submodule_kineto 2025-12-04T08:53:58.7838838Z * [new branch] update_submodule_tensorpipe -> origin/update_submodule_tensorpipe 2025-12-04T08:53:58.7838977Z * [new branch] upload-tests-for-autorevert -> origin/upload-tests-for-autorevert 2025-12-04T08:53:58.7839041Z * [new branch] v0.1.2 -> origin/v0.1.2 2025-12-04T08:53:58.7839102Z * [new branch] v1.0.1 -> origin/v1.0.1 2025-12-04T08:53:58.7839161Z * [new branch] v1.0.3 -> origin/v1.0.3 2025-12-04T08:53:58.7839219Z * [new branch] v1.1.0 -> origin/v1.1.0 2025-12-04T08:53:58.7839277Z * [new branch] v1.2.0 -> origin/v1.2.0 2025-12-04T08:53:58.7839335Z * [new branch] v1.3.0 -> origin/v1.3.0 2025-12-04T08:53:58.7839393Z * [new branch] v1.3.1 -> origin/v1.3.1 2025-12-04T08:53:58.7839457Z * [new branch] validate_fn -> origin/validate_fn 2025-12-04T08:53:58.7839526Z * [new branch] validations_2.6 -> origin/validations_2.6 2025-12-04T08:53:58.7839595Z * [new branch] validations_2.8 -> origin/validations_2.8 2025-12-04T08:53:58.7839660Z * [new branch] varlen-api -> origin/varlen-api 2025-12-04T08:53:58.7839735Z * [new branch] varlen-api-backup -> origin/varlen-api-backup 2025-12-04T08:53:58.7839814Z * [new branch] varlen_batch_invariance -> origin/varlen_batch_invariance 2025-12-04T08:53:58.7839879Z * [new branch] viable/strict -> origin/viable/strict 2025-12-04T08:53:58.7839999Z * [new branch] vishal9-team/dtensor_parallelism_toy -> origin/vishal9-team/dtensor_parallelism_toy 2025-12-04T08:53:58.7840064Z * [new branch] vllmbuildci -> origin/vllmbuildci 2025-12-04T08:53:58.7840125Z * [new branch] vllmpin -> origin/vllmpin 2025-12-04T08:53:58.7840214Z * [new branch] vscode-recommend-pyrefly -> origin/vscode-recommend-pyrefly 2025-12-04T08:53:58.7840284Z * [new branch] wdvr-patch-1 -> origin/wdvr-patch-1 2025-12-04T08:53:58.7840349Z * [new branch] wdvr/iss_145259 -> origin/wdvr/iss_145259 2025-12-04T08:53:58.7840411Z * [new branch] whc/pei -> origin/whc/pei 2025-12-04T08:53:58.7840477Z * [new branch] whc/pp_fix -> origin/whc/pp_fix 2025-12-04T08:53:58.7840541Z * [new branch] whc/sharding -> origin/whc/sharding 2025-12-04T08:53:58.7840605Z * [new branch] whc/sharding2 -> origin/whc/sharding2 2025-12-04T08:53:58.7840704Z * [new branch] whc/uneven -> origin/whc/uneven 2025-12-04T08:53:58.7840775Z * [new branch] whc/uneven-merge -> origin/whc/uneven-merge 2025-12-04T08:53:58.7840838Z * [new branch] win_warnings -> origin/win_warnings 2025-12-04T08:53:58.7840915Z * [new branch] windows_libtorch_free -> origin/windows_libtorch_free 2025-12-04T08:53:58.7840977Z * [new branch] xmfan-war -> origin/xmfan-war 2025-12-04T08:53:58.7841045Z * [new branch] xmfan/ca_0516 -> origin/xmfan/ca_0516 2025-12-04T08:53:58.7841113Z * [new branch] xmfan/ca_1051b93192 -> origin/xmfan/ca_1051b93192 2025-12-04T08:53:58.7841264Z * [new branch] xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 -> origin/xmfan/ca_1a722f62c248391fc4a542e8851a5559aa356ae8 2025-12-04T08:53:58.7841336Z * [new branch] xmfan/ca_5a2be192d1 -> origin/xmfan/ca_5a2be192d1 2025-12-04T08:53:58.7841406Z * [new branch] xmfan/ca_9d59b516e9 -> origin/xmfan/ca_9d59b516e9 2025-12-04T08:53:58.7841471Z * [new branch] xmfan/ca_apr8 -> origin/xmfan/ca_apr8 2025-12-04T08:53:58.7841537Z * [new branch] xmfan/ca_base -> origin/xmfan/ca_base 2025-12-04T08:53:58.7841644Z * [new branch] xmfan/ca_dynamic -> origin/xmfan/ca_dynamic 2025-12-04T08:53:58.7841712Z * [new branch] xmfan/ca_fix_dyn -> origin/xmfan/ca_fix_dyn 2025-12-04T08:53:58.7841787Z * [new branch] xmfan/ca_fix_lowering -> origin/xmfan/ca_fix_lowering 2025-12-04T08:53:58.7841863Z * [new branch] xmfan/ca_fix_polyfills -> origin/xmfan/ca_fix_polyfills 2025-12-04T08:53:58.7841926Z * [new branch] xmfan/ca_jan3 -> origin/xmfan/ca_jan3 2025-12-04T08:53:58.7841993Z * [new branch] xmfan/ca_jun18 -> origin/xmfan/ca_jun18 2025-12-04T08:53:58.7842059Z * [new branch] xmfan/ca_jun24 -> origin/xmfan/ca_jun24 2025-12-04T08:53:58.7842125Z * [new branch] xmfan/ca_nested -> origin/xmfan/ca_nested 2025-12-04T08:53:58.7842194Z * [new branch] xmfan/ca_overhead -> origin/xmfan/ca_overhead 2025-12-04T08:53:58.7842288Z * [new branch] xmfan/ca_overhead_0eba7e5451 -> origin/xmfan/ca_overhead_0eba7e5451 2025-12-04T08:53:58.7842355Z * [new branch] xmfan/cacu_jun18 -> origin/xmfan/cacu_jun18 2025-12-04T08:53:58.7842423Z * [new branch] xmfan/cacu_jun19 -> origin/xmfan/cacu_jun19 2025-12-04T08:53:58.7842488Z * [new branch] xmfan/cacu_jun4 -> origin/xmfan/cacu_jun4 2025-12-04T08:53:58.7842571Z * [new branch] xmfan/disable_duck_shape -> origin/xmfan/disable_duck_shape 2025-12-04T08:53:58.7842668Z * [new branch] xmfan/fca_cpp_node_passthrough -> origin/xmfan/fca_cpp_node_passthrough 2025-12-04T08:53:58.7842824Z * [new branch] xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/post_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T08:53:58.7842973Z * [new branch] xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 -> origin/xmfan/pre_3945954741e2d37023c5d6954f9483008e0892f9 2025-12-04T08:53:58.7843046Z * [new branch] xmfan/single_step -> origin/xmfan/single_step 2025-12-04T08:53:58.7843111Z * [new branch] xmfan/sth_0829 -> origin/xmfan/sth_0829 2025-12-04T08:53:58.7843174Z * [new branch] xmfan/test -> origin/xmfan/test 2025-12-04T08:53:58.7843263Z * [new branch] yguo/debug-0226-constexpr -> origin/yguo/debug-0226-constexpr 2025-12-04T08:53:58.7843340Z * [new branch] yguo/new_latest_changes -> origin/yguo/new_latest_changes 2025-12-04T08:53:58.7843436Z * [new branch] yguo/patch_constexpr_changes -> origin/yguo/patch_constexpr_changes 2025-12-04T08:53:58.7843531Z * [new branch] yiming/bootcamp -> origin/yiming/bootcamp 2025-12-04T08:53:58.7843634Z * [new branch] yiming/run_with_start_end_rng_hop -> origin/yiming/run_with_start_end_rng_hop 2025-12-04T08:53:58.7843700Z * [new branch] yolo-llama3 -> origin/yolo-llama3 2025-12-04T08:53:58.7843773Z * [new branch] zainr/canary-test -> origin/zainr/canary-test 2025-12-04T08:53:58.7843860Z * [new branch] zainr/cleanup-gh-runners -> origin/zainr/cleanup-gh-runners 2025-12-04T08:53:58.7843941Z * [new branch] zainr/pull-migration-c -> origin/zainr/pull-migration-c 2025-12-04T08:53:58.7844004Z * [new branch] zainr/test2 -> origin/zainr/test2 2025-12-04T08:53:58.7844078Z * [new branch] zasdfgbnm-patch-3 -> origin/zasdfgbnm-patch-3 2025-12-04T08:53:58.7844139Z * [new branch] zb2p -> origin/zb2p 2025-12-04T08:53:58.7844225Z * [new branch] zeros-and-scatter-part2 -> origin/zeros-and-scatter-part2 2025-12-04T08:53:58.7844313Z * [new branch] zhxchen17/ci/vllm_lora_oom -> origin/zhxchen17/ci/vllm_lora_oom 2025-12-04T08:53:58.7844416Z * [new branch] zhxchen17/ci/vllm_multimodal_oom -> origin/zhxchen17/ci/vllm_multimodal_oom 2025-12-04T08:53:58.7844519Z * [new branch] zhxchen17/ci/vllm_pin -> origin/zhxchen17/ci/vllm_pin 2025-12-04T08:53:58.7844643Z * [new branch] zhxchen17/dynamo/unsafe_drop_all_guards -> origin/zhxchen17/dynamo/unsafe_drop_all_guards 2025-12-04T08:53:58.7844741Z * [new branch] zhxchen17/export/call_override -> origin/zhxchen17/export/call_override 2025-12-04T08:53:58.7844828Z * [new branch] zhxchen17/export/codemod1 -> origin/zhxchen17/export/codemod1 2025-12-04T08:53:58.7844918Z * [new branch] zhxchen17/export/ctx_return -> origin/zhxchen17/export/ctx_return 2025-12-04T08:53:58.7845047Z * [new branch] zhxchen17/export/disable_side_effect_warn -> origin/zhxchen17/export/disable_side_effect_warn 2025-12-04T08:53:58.7845146Z * [new branch] zhxchen17/export/pytree_check -> origin/zhxchen17/export/pytree_check 2025-12-04T08:53:58.7845235Z * [new branch] zhxchen17/precompile/aoti -> origin/zhxchen17/precompile/aoti 2025-12-04T08:53:58.7845332Z * [new branch] zhxchen17/precompile/globals -> origin/zhxchen17/precompile/globals 2025-12-04T08:53:58.7845449Z * [new branch] zhxchen17/precompile/inductor_guards -> origin/zhxchen17/precompile/inductor_guards 2025-12-04T08:53:58.7845524Z * [new branch] zhxchen17/scratch/0 -> origin/zhxchen17/scratch/0 2025-12-04T08:53:58.7845631Z * [new branch] zhxchen17/torch_export_api_update -> origin/zhxchen17/torch_export_api_update 2025-12-04T08:53:58.7845708Z * [new branch] zhxhcen17/moodycamel -> origin/zhxhcen17/moodycamel 2025-12-04T08:53:58.7845783Z * [new branch] zxiiro/build-times -> origin/zxiiro/build-times 2025-12-04T08:53:58.7845856Z * [new branch] zxiiro/c7i.2xlarge -> origin/zxiiro/c7i.2xlarge 2025-12-04T08:53:58.7845935Z * [new branch] zxiiro/c7i.2xlarge.h100 -> origin/zxiiro/c7i.2xlarge.h100 2025-12-04T08:53:58.7845999Z * [new branch] zxiiro/main -> origin/zxiiro/main 2025-12-04T08:53:58.7846064Z * [new branch] zxiiro/risc64 -> origin/zxiiro/risc64 2025-12-04T08:53:58.7846156Z * [new branch] zxiiro/test-multicloud-arc -> origin/zxiiro/test-multicloud-arc 2025-12-04T08:53:58.7846216Z * [new tag] ciflow/dynamo/169525 -> ciflow/dynamo/169525 2025-12-04T08:53:58.7846287Z t [tag update] ciflow/inductor/167647 -> ciflow/inductor/167647 2025-12-04T08:53:58.7846380Z t [tag update] ciflow/inductor/168266 -> ciflow/inductor/168266 2025-12-04T08:53:58.7846448Z t [tag update] ciflow/inductor/169535 -> ciflow/inductor/169535 2025-12-04T08:53:58.7846510Z * [new tag] ciflow/trunk/165728 -> ciflow/trunk/165728 2025-12-04T08:53:58.7846570Z * [new tag] ciflow/trunk/169048 -> ciflow/trunk/169048 2025-12-04T08:53:58.7846631Z * [new tag] ciflow/trunk/169125 -> ciflow/trunk/169125 2025-12-04T08:53:58.7846690Z * [new tag] ciflow/trunk/169555 -> ciflow/trunk/169555 2025-12-04T08:53:58.7846803Z * [new tag] ciflow/xpu/169555 -> ciflow/xpu/169555 2025-12-04T08:53:58.9720409Z [command]/usr/bin/git rev-parse --verify --quiet ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32^{object} 2025-12-04T08:53:58.9858365Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:53:58.9862664Z ##[endgroup] 2025-12-04T08:53:58.9862806Z ##[group]Determining the checkout info 2025-12-04T08:53:58.9864909Z ##[endgroup] 2025-12-04T08:53:58.9869829Z [command]/usr/bin/git sparse-checkout disable 2025-12-04T08:53:58.9956351Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-12-04T08:53:58.9971760Z ##[group]Checking out the ref 2025-12-04T08:53:58.9973259Z [command]/usr/bin/git checkout --progress --force ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:53:59.0218403Z HEAD is now at ffd9b0fb4355 Resolve collective autotuning test failure on arm (#168919) 2025-12-04T08:53:59.0224258Z ##[endgroup] 2025-12-04T08:53:59.0224434Z ##[group]Setting up auth for fetching submodules 2025-12-04T08:53:59.0230190Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T08:53:59.0256058Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-12-04T08:53:59.0278082Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-12-04T08:53:59.0299602Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-12-04T08:53:59.0322408Z ##[endgroup] 2025-12-04T08:53:59.0322605Z ##[group]Fetching submodules 2025-12-04T08:53:59.0324274Z [command]/usr/bin/git submodule sync --recursive 2025-12-04T08:53:59.0516081Z Synchronizing submodule url for 'android/libs/fbjni' 2025-12-04T08:53:59.0527151Z Synchronizing submodule url for 'third_party/FP16' 2025-12-04T08:53:59.0540154Z Synchronizing submodule url for 'third_party/FXdiv' 2025-12-04T08:53:59.0556335Z Synchronizing submodule url for 'third_party/NNPACK' 2025-12-04T08:53:59.0566714Z Synchronizing submodule url for 'third_party/NVTX' 2025-12-04T08:53:59.0579880Z Synchronizing submodule url for 'third_party/VulkanMemoryAllocator' 2025-12-04T08:53:59.0596913Z Synchronizing submodule url for 'third_party/XNNPACK' 2025-12-04T08:53:59.0614220Z Synchronizing submodule url for 'third_party/aiter' 2025-12-04T08:53:59.0634361Z Synchronizing submodule url for 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:53:59.0649635Z Synchronizing submodule url for 'third_party/benchmark' 2025-12-04T08:53:59.0661336Z Synchronizing submodule url for 'third_party/composable_kernel' 2025-12-04T08:53:59.0677723Z Synchronizing submodule url for 'third_party/cpp-httplib' 2025-12-04T08:53:59.0694855Z Synchronizing submodule url for 'third_party/cpuinfo' 2025-12-04T08:53:59.0708157Z Synchronizing submodule url for 'third_party/cudnn_frontend' 2025-12-04T08:53:59.0719231Z Synchronizing submodule url for 'third_party/cutlass' 2025-12-04T08:53:59.0734874Z Synchronizing submodule url for 'third_party/fbgemm' 2025-12-04T08:53:59.0749308Z Synchronizing submodule url for 'third_party/fbgemm/external/asmjit' 2025-12-04T08:53:59.0761103Z Synchronizing submodule url for 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:53:59.0781235Z Synchronizing submodule url for 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:53:59.0794482Z Synchronizing submodule url for 'third_party/fbgemm/external/cutlass' 2025-12-04T08:53:59.0809020Z Synchronizing submodule url for 'third_party/fbgemm/external/googletest' 2025-12-04T08:53:59.0819672Z Synchronizing submodule url for 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:53:59.0830023Z Synchronizing submodule url for 'third_party/fbgemm/external/json' 2025-12-04T08:53:59.0843500Z Synchronizing submodule url for 'third_party/flash-attention' 2025-12-04T08:53:59.0857794Z Synchronizing submodule url for 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:53:59.0868947Z Synchronizing submodule url for 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:53:59.0886680Z Synchronizing submodule url for 'third_party/flatbuffers' 2025-12-04T08:53:59.0900068Z Synchronizing submodule url for 'third_party/fmt' 2025-12-04T08:53:59.0911449Z Synchronizing submodule url for 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:53:59.0929800Z Synchronizing submodule url for 'third_party/gloo' 2025-12-04T08:53:59.0941752Z Synchronizing submodule url for 'third_party/googletest' 2025-12-04T08:53:59.0955404Z Synchronizing submodule url for 'third_party/ideep' 2025-12-04T08:53:59.0967263Z Synchronizing submodule url for 'third_party/ideep/mkl-dnn' 2025-12-04T08:53:59.0981909Z Synchronizing submodule url for 'third_party/ittapi' 2025-12-04T08:53:59.0993465Z Synchronizing submodule url for 'third_party/kineto' 2025-12-04T08:53:59.1006205Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:53:59.1020385Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:53:59.1031322Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:53:59.1042668Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:53:59.1053267Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:53:59.1065190Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:53:59.1076725Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:53:59.1089189Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:53:59.1100065Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:53:59.1114013Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:53:59.1131966Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:53:59.1143226Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:53:59.1154320Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:53:59.1167801Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:53:59.1177939Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:53:59.1194716Z Synchronizing submodule url for 'third_party/kleidiai' 2025-12-04T08:53:59.1207085Z Synchronizing submodule url for 'third_party/mimalloc' 2025-12-04T08:53:59.1223483Z Synchronizing submodule url for 'third_party/nlohmann' 2025-12-04T08:53:59.1235446Z Synchronizing submodule url for 'third_party/onnx' 2025-12-04T08:53:59.1254653Z Synchronizing submodule url for 'third_party/onnx/third_party/pybind11' 2025-12-04T08:53:59.1273820Z Synchronizing submodule url for 'third_party/opentelemetry-cpp' 2025-12-04T08:53:59.1288385Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:53:59.1301409Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:53:59.1312356Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:53:59.1327555Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:53:59.1339623Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:53:59.1349089Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:53:59.1358193Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:53:59.1370508Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:53:59.1384950Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:53:59.1396256Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:53:59.1418608Z Synchronizing submodule url for 'third_party/pocketfft' 2025-12-04T08:53:59.1432331Z Synchronizing submodule url for 'third_party/protobuf' 2025-12-04T08:53:59.1447901Z Synchronizing submodule url for 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:53:59.1457842Z Synchronizing submodule url for 'third_party/protobuf/third_party/googletest' 2025-12-04T08:53:59.1470199Z Synchronizing submodule url for 'third_party/psimd' 2025-12-04T08:53:59.1486698Z Synchronizing submodule url for 'third_party/pthreadpool' 2025-12-04T08:53:59.1499838Z Synchronizing submodule url for 'third_party/pybind11' 2025-12-04T08:53:59.1511809Z Synchronizing submodule url for 'third_party/python-peachpy' 2025-12-04T08:53:59.1523061Z Synchronizing submodule url for 'third_party/sleef' 2025-12-04T08:53:59.1532765Z Synchronizing submodule url for 'third_party/tensorpipe' 2025-12-04T08:53:59.1543099Z Synchronizing submodule url for 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:53:59.1554627Z Synchronizing submodule url for 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:53:59.1566473Z Synchronizing submodule url for 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:53:59.1577754Z Synchronizing submodule url for 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:53:59.1589955Z Synchronizing submodule url for 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:53:59.1613803Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-12-04T08:53:59.1847286Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-12-04T08:53:59.1910770Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-12-04T08:53:59.1955782Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-12-04T08:53:59.2070030Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-12-04T08:53:59.2131714Z Submodule path 'third_party/NVTX': checked out '3ebbc93ded7285963bff932c678fa367eb393ba6' 2025-12-04T08:53:59.2183824Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-12-04T08:53:59.6961414Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-12-04T08:53:59.7114297Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-12-04T08:53:59.7296306Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-12-04T08:53:59.7410812Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-12-04T08:53:59.7587715Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T08:53:59.7659003Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-12-04T08:53:59.8292358Z Submodule path 'third_party/cpuinfo': checked out 'f858c30bcb16f8effd5ff46996f0514539e17abc' 2025-12-04T08:53:59.8374921Z Submodule path 'third_party/cudnn_frontend': checked out '0b1577c8c83401237d601d0d0db5210506705396' 2025-12-04T08:53:59.8494784Z Submodule path 'third_party/cutlass': checked out 'f88806b1e31dfa579842638740216dd41fc6c588' 2025-12-04T08:53:59.9178085Z Submodule path 'third_party/fbgemm': checked out 'c0b988d39a9e47c794d699f29930ed4d7c7e13a4' 2025-12-04T08:53:59.9492023Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-12-04T08:54:00.1227633Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T08:54:00.1861895Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-12-04T08:54:00.6188416Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '98125ce499b0fdf7ffbe0e3052f5b8709f4840f8' 2025-12-04T08:54:00.6397247Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:54:00.6496262Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-12-04T08:54:00.7043547Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-12-04T08:54:00.7148001Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-12-04T08:54:00.7327650Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-12-04T08:54:00.7432953Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-12-04T08:54:00.7524886Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-12-04T08:54:00.7665208Z Submodule path 'third_party/fmt': checked out '407c905e45ad75fc29bf0f9bb7c5c2fd3475976f' 2025-12-04T08:54:00.7871088Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-12-04T08:54:00.7980241Z Submodule path 'third_party/gloo': checked out '54cbae0d3a67fa890b4c3d9ee162b7860315e341' 2025-12-04T08:54:00.8153635Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:54:00.8234668Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-12-04T08:54:01.2198202Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-12-04T08:54:01.2295702Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-12-04T08:54:01.2384903Z Submodule path 'third_party/kineto': checked out '31f85df8fbd89c188f14ef10f1ec65379786b943' 2025-12-04T08:54:01.2471936Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out 'd2ffe0a4e3acace628db49974246b66fc3e85fb1' 2025-12-04T08:54:01.2566248Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-12-04T08:54:01.2640093Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-12-04T08:54:01.2710992Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-12-04T08:54:01.2778636Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-12-04T08:54:01.2830287Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-12-04T08:54:01.2879683Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-12-04T08:54:01.2963440Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:54:01.3052729Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-12-04T08:54:01.3115168Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-12-04T08:54:01.3176192Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp': checked out 'b1234816facfdda29845c46696a02998a4af115a' 2025-12-04T08:54:01.3250558Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'd7ba35bbb649209c66e582d5a0244ba988a15159' 2025-12-04T08:54:01.3308895Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T08:54:01.3394142Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-12-04T08:54:01.3454574Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:54:01.3539316Z Submodule path 'third_party/kleidiai': checked out 'd7770c89632329a9914ef1a90289917597639cbe' 2025-12-04T08:54:01.3613284Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-12-04T08:54:01.3706050Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-12-04T08:54:01.5493503Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-12-04T08:54:01.5699072Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-12-04T08:54:01.5811409Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-12-04T08:54:01.5902160Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-12-04T08:54:01.5966888Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-12-04T08:54:01.6027853Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-12-04T08:54:01.6122329Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-12-04T08:54:01.6178756Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-12-04T08:54:01.6225855Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-12-04T08:54:01.6291551Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-12-04T08:54:01.6366719Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-12-04T08:54:01.6431223Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T08:54:01.6589488Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-12-04T08:54:01.6652834Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-12-04T08:54:01.6961943Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-12-04T08:54:01.7044896Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-12-04T08:54:01.7256187Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-12-04T08:54:01.7314570Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-12-04T08:54:01.7389634Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-12-04T08:54:01.7577311Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-12-04T08:54:01.7796457Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-12-04T08:54:01.8044054Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-12-04T08:54:01.8165109Z Submodule path 'third_party/tensorpipe': checked out '2b4cd91092d335a697416b2a3cb398283246849d' 2025-12-04T08:54:01.8354754Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-12-04T08:54:01.8443388Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-12-04T08:54:01.8735630Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-12-04T08:54:01.8879353Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-12-04T08:54:01.8944739Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-12-04T08:54:01.8978399Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-12-04T08:54:01.9165248Z Entering 'android/libs/fbjni' 2025-12-04T08:54:01.9190732Z Entering 'third_party/FP16' 2025-12-04T08:54:01.9210699Z Entering 'third_party/FXdiv' 2025-12-04T08:54:01.9229178Z Entering 'third_party/NNPACK' 2025-12-04T08:54:01.9250398Z Entering 'third_party/NVTX' 2025-12-04T08:54:01.9269991Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:01.9291339Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:01.9320966Z Entering 'third_party/aiter' 2025-12-04T08:54:01.9343462Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:01.9370933Z Entering 'third_party/benchmark' 2025-12-04T08:54:01.9402236Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:01.9424574Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:01.9445131Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:01.9466185Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:01.9485080Z Entering 'third_party/cutlass' 2025-12-04T08:54:01.9506935Z Entering 'third_party/fbgemm' 2025-12-04T08:54:01.9527663Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:01.9547283Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:01.9572389Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:01.9598801Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:01.9625344Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:01.9648615Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:01.9668432Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:01.9688247Z Entering 'third_party/flash-attention' 2025-12-04T08:54:01.9709017Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:01.9731293Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:01.9761151Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:01.9784211Z Entering 'third_party/fmt' 2025-12-04T08:54:01.9803314Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:01.9828354Z Entering 'third_party/gloo' 2025-12-04T08:54:01.9847263Z Entering 'third_party/googletest' 2025-12-04T08:54:01.9865628Z Entering 'third_party/ideep' 2025-12-04T08:54:01.9891279Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:01.9927025Z Entering 'third_party/ittapi' 2025-12-04T08:54:01.9948493Z Entering 'third_party/kineto' 2025-12-04T08:54:01.9969637Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:01.9990748Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:02.0020931Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:02.0041005Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:02.0059224Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:02.0094207Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:02.0122042Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:02.0143730Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:02.0173694Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:02.0193395Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:02.0212265Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:02.0237125Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:02.0256117Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:02.0277646Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:02.0294959Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:02.0314153Z Entering 'third_party/kleidiai' 2025-12-04T08:54:02.0341376Z Entering 'third_party/mimalloc' 2025-12-04T08:54:02.0369117Z Entering 'third_party/nlohmann' 2025-12-04T08:54:02.0389551Z Entering 'third_party/onnx' 2025-12-04T08:54:02.0413792Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:02.0442158Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:02.0461930Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:02.0479388Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:02.0496389Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:02.0514289Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:02.0538112Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:02.0565001Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:02.0589396Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:02.0608731Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:02.0635704Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:02.0662899Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:02.0703847Z Entering 'third_party/pocketfft' 2025-12-04T08:54:02.0727182Z Entering 'third_party/protobuf' 2025-12-04T08:54:02.0756582Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:02.0773006Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:02.0795442Z Entering 'third_party/psimd' 2025-12-04T08:54:02.0820282Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:02.0840686Z Entering 'third_party/pybind11' 2025-12-04T08:54:02.0864587Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:02.0882854Z Entering 'third_party/sleef' 2025-12-04T08:54:02.0902045Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:02.0926347Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:02.0948925Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:02.0967990Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:02.0986404Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:02.1006589Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:02.1036705Z ##[endgroup] 2025-12-04T08:54:02.1036917Z ##[group]Persisting credentials for submodules 2025-12-04T08:54:02.1046139Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-12-04T08:54:02.1214613Z Entering 'android/libs/fbjni' 2025-12-04T08:54:02.1236286Z Entering 'third_party/FP16' 2025-12-04T08:54:02.1265894Z Entering 'third_party/FXdiv' 2025-12-04T08:54:02.1292851Z Entering 'third_party/NNPACK' 2025-12-04T08:54:02.1316105Z Entering 'third_party/NVTX' 2025-12-04T08:54:02.1337870Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:02.1358550Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:02.1403699Z Entering 'third_party/aiter' 2025-12-04T08:54:02.1427472Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:02.1455452Z Entering 'third_party/benchmark' 2025-12-04T08:54:02.1478608Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:02.1507251Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:02.1529976Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:02.1551379Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:02.1574983Z Entering 'third_party/cutlass' 2025-12-04T08:54:02.1600978Z Entering 'third_party/fbgemm' 2025-12-04T08:54:02.1623172Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:02.1645327Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:02.1669399Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:02.1694006Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:02.1720182Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:02.1747679Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:02.1770214Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:02.1792041Z Entering 'third_party/flash-attention' 2025-12-04T08:54:02.1820933Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:02.1847011Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:02.1874232Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:02.1897861Z Entering 'third_party/fmt' 2025-12-04T08:54:02.1921120Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:02.1948239Z Entering 'third_party/gloo' 2025-12-04T08:54:02.1974668Z Entering 'third_party/googletest' 2025-12-04T08:54:02.1996825Z Entering 'third_party/ideep' 2025-12-04T08:54:02.2018816Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:02.2053665Z Entering 'third_party/ittapi' 2025-12-04T08:54:02.2074091Z Entering 'third_party/kineto' 2025-12-04T08:54:02.2095232Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:02.2117608Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:02.2141254Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:02.2168894Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:02.2192210Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:02.2214552Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:02.2236597Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:02.2255745Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:02.2276957Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:02.2298079Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:02.2318435Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:02.2340242Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:02.2362008Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:02.2388591Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:02.2412589Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:02.2441162Z Entering 'third_party/kleidiai' 2025-12-04T08:54:02.2462912Z Entering 'third_party/mimalloc' 2025-12-04T08:54:02.2484013Z Entering 'third_party/nlohmann' 2025-12-04T08:54:02.2509422Z Entering 'third_party/onnx' 2025-12-04T08:54:02.2538007Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:02.2563195Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:02.2586321Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:02.2609104Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:02.2632545Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:02.2667348Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:02.2689415Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:02.2710300Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:02.2733247Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:02.2756545Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:02.2781562Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:02.2805465Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:02.2837131Z Entering 'third_party/pocketfft' 2025-12-04T08:54:02.2860701Z Entering 'third_party/protobuf' 2025-12-04T08:54:02.2888102Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:02.2912713Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:02.2936867Z Entering 'third_party/psimd' 2025-12-04T08:54:02.2958182Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:02.2987015Z Entering 'third_party/pybind11' 2025-12-04T08:54:02.3010230Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:02.3032910Z Entering 'third_party/sleef' 2025-12-04T08:54:02.3053222Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:02.3074595Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:02.3097420Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:02.3118679Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:02.3139468Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:02.3160785Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:02.3204221Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-12-04T08:54:02.3377511Z Entering 'android/libs/fbjni' 2025-12-04T08:54:02.3404986Z file:/home/runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T08:54:02.3415126Z Entering 'third_party/FP16' 2025-12-04T08:54:02.3439292Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T08:54:02.3449393Z Entering 'third_party/FXdiv' 2025-12-04T08:54:02.3477170Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T08:54:02.3489077Z Entering 'third_party/NNPACK' 2025-12-04T08:54:02.3509507Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T08:54:02.3519546Z Entering 'third_party/NVTX' 2025-12-04T08:54:02.3539524Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T08:54:02.3548606Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:02.3566701Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T08:54:02.3575974Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:02.3595226Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T08:54:02.3608656Z Entering 'third_party/aiter' 2025-12-04T08:54:02.3625950Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T08:54:02.3636072Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:02.3658384Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T08:54:02.3673098Z Entering 'third_party/benchmark' 2025-12-04T08:54:02.3692447Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:54:02.3707249Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:02.3729307Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T08:54:02.3748835Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:02.3769770Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T08:54:02.3779649Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:02.3802528Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T08:54:02.3812651Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:02.3832931Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T08:54:02.3843737Z Entering 'third_party/cutlass' 2025-12-04T08:54:02.3867195Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T08:54:02.3880852Z Entering 'third_party/fbgemm' 2025-12-04T08:54:02.3901691Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T08:54:02.3912115Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:02.3931081Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T08:54:02.3943389Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:02.3963728Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T08:54:02.3976127Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:02.3998553Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T08:54:02.4007735Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:02.4028536Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T08:54:02.4042130Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:02.4063981Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T08:54:02.4081260Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:02.4099649Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T08:54:02.4108692Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:02.4127886Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T08:54:02.4139457Z Entering 'third_party/flash-attention' 2025-12-04T08:54:02.4156525Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T08:54:02.4164910Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:02.4188995Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T08:54:02.4200355Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:02.4246064Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T08:54:02.4265269Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:02.4299652Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T08:54:02.4309524Z Entering 'third_party/fmt' 2025-12-04T08:54:02.4329635Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T08:54:02.4340120Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:02.4361845Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T08:54:02.4371101Z Entering 'third_party/gloo' 2025-12-04T08:54:02.4389450Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T08:54:02.4402136Z Entering 'third_party/googletest' 2025-12-04T08:54:02.4420749Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:02.4433296Z Entering 'third_party/ideep' 2025-12-04T08:54:02.4457436Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T08:54:02.4467606Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:02.4485906Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T08:54:02.4498827Z Entering 'third_party/ittapi' 2025-12-04T08:54:02.4518294Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T08:54:02.4528118Z Entering 'third_party/kineto' 2025-12-04T08:54:02.4552502Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T08:54:02.4562485Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:02.4597582Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T08:54:02.4608201Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:02.4641767Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T08:54:02.4652709Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:02.4688656Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T08:54:02.4699771Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:02.4731108Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T08:54:02.4743392Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:02.4766886Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T08:54:02.4774739Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:02.4795712Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T08:54:02.4808662Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:02.4847162Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T08:54:02.4856209Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:02.4881909Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:02.4897483Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:02.4920784Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T08:54:02.4930925Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:02.4951882Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T08:54:02.4962606Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:02.4982763Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T08:54:02.4995376Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:02.5015923Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T08:54:02.5026508Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:02.5051548Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T08:54:02.5063077Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:02.5085362Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T08:54:02.5100762Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:02.5123397Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T08:54:02.5133202Z Entering 'third_party/kleidiai' 2025-12-04T08:54:02.5152047Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T08:54:02.5161733Z Entering 'third_party/mimalloc' 2025-12-04T08:54:02.5179725Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T08:54:02.5193378Z Entering 'third_party/nlohmann' 2025-12-04T08:54:02.5218372Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T08:54:02.5228997Z Entering 'third_party/onnx' 2025-12-04T08:54:02.5250685Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T08:54:02.5275942Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:02.5312451Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:54:02.5329561Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:02.5354136Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T08:54:02.5363731Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:02.5390121Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:54:02.5400617Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:02.5425872Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:02.5437421Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:02.5461765Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T08:54:02.5471537Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:02.5495098Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T08:54:02.5505119Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:02.5523748Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T08:54:02.5532898Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:02.5553075Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T08:54:02.5561226Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:02.5579446Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T08:54:02.5593636Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:02.5615106Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T08:54:02.5625084Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:02.5647973Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T08:54:02.5659239Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:02.5678055Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T08:54:02.5695820Z Entering 'third_party/pocketfft' 2025-12-04T08:54:02.5714696Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T08:54:02.5723093Z Entering 'third_party/protobuf' 2025-12-04T08:54:02.5743783Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T08:54:02.5755330Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:02.5778817Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:54:02.5788209Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:02.5809853Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:02.5821000Z Entering 'third_party/psimd' 2025-12-04T08:54:02.5843617Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T08:54:02.5852238Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:02.5871621Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T08:54:02.5880024Z Entering 'third_party/pybind11' 2025-12-04T08:54:02.5904815Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:54:02.5914689Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:02.5933260Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T08:54:02.5942494Z Entering 'third_party/sleef' 2025-12-04T08:54:02.5963924Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T08:54:02.5973208Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:02.5991333Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T08:54:02.6001017Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:02.6029128Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:02.6038932Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:02.6064314Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T08:54:02.6072844Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:02.6091411Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T08:54:02.6100340Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:02.6119824Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:54:02.6128649Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:02.6148472Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T08:54:02.6322165Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-12-04T08:54:02.6492286Z Entering 'android/libs/fbjni' 2025-12-04T08:54:02.6513993Z Entering 'third_party/FP16' 2025-12-04T08:54:02.6535976Z Entering 'third_party/FXdiv' 2025-12-04T08:54:02.6555864Z Entering 'third_party/NNPACK' 2025-12-04T08:54:02.6579042Z Entering 'third_party/NVTX' 2025-12-04T08:54:02.6602794Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:02.6626833Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:02.6650977Z Entering 'third_party/aiter' 2025-12-04T08:54:02.6672122Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:02.6703827Z Entering 'third_party/benchmark' 2025-12-04T08:54:02.6723897Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:02.6745445Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:02.6763928Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:02.6783553Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:02.6809886Z Entering 'third_party/cutlass' 2025-12-04T08:54:02.6837827Z Entering 'third_party/fbgemm' 2025-12-04T08:54:02.6857961Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:02.6876611Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:02.6896884Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:02.6916469Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:02.6940468Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:02.6960392Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:02.6977795Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:02.7000092Z Entering 'third_party/flash-attention' 2025-12-04T08:54:02.7018960Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:02.7040968Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:02.7063103Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:02.7085326Z Entering 'third_party/fmt' 2025-12-04T08:54:02.7104422Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:02.7124258Z Entering 'third_party/gloo' 2025-12-04T08:54:02.7142185Z Entering 'third_party/googletest' 2025-12-04T08:54:02.7166603Z Entering 'third_party/ideep' 2025-12-04T08:54:02.7188740Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:02.7209858Z Entering 'third_party/ittapi' 2025-12-04T08:54:02.7234345Z Entering 'third_party/kineto' 2025-12-04T08:54:02.7254769Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:02.7281070Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:02.7299095Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:02.7318501Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:02.7336466Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:02.7353392Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:02.7372242Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:02.7391370Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:02.7414174Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:02.7440446Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:02.7459592Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:02.7486098Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:02.7511358Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:02.7534445Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:02.7552406Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:02.7572620Z Entering 'third_party/kleidiai' 2025-12-04T08:54:02.7590733Z Entering 'third_party/mimalloc' 2025-12-04T08:54:02.7610401Z Entering 'third_party/nlohmann' 2025-12-04T08:54:02.7629591Z Entering 'third_party/onnx' 2025-12-04T08:54:02.7654362Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:02.7680501Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:02.7699207Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:02.7723324Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:02.7743148Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:02.7760545Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:02.7779735Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:02.7797590Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:02.7815222Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:02.7832180Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:02.7851097Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:02.7871856Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:02.7901758Z Entering 'third_party/pocketfft' 2025-12-04T08:54:02.7927292Z Entering 'third_party/protobuf' 2025-12-04T08:54:02.7948113Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:02.7965816Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:02.7987650Z Entering 'third_party/psimd' 2025-12-04T08:54:02.8006418Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:02.8024757Z Entering 'third_party/pybind11' 2025-12-04T08:54:02.8048171Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:02.8066483Z Entering 'third_party/sleef' 2025-12-04T08:54:02.8085628Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:02.8107440Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:02.8130886Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:02.8150043Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:02.8169677Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:02.8189340Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:02.8221799Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-12-04T08:54:02.8386279Z Entering 'android/libs/fbjni' 2025-12-04T08:54:02.8406918Z Entering 'third_party/FP16' 2025-12-04T08:54:02.8424575Z Entering 'third_party/FXdiv' 2025-12-04T08:54:02.8443087Z Entering 'third_party/NNPACK' 2025-12-04T08:54:02.8464375Z Entering 'third_party/NVTX' 2025-12-04T08:54:02.8486675Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:02.8505683Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:02.8532979Z Entering 'third_party/aiter' 2025-12-04T08:54:02.8555285Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:02.8581848Z Entering 'third_party/benchmark' 2025-12-04T08:54:02.8601262Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:02.8622834Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:02.8640855Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:02.8658981Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:02.8682607Z Entering 'third_party/cutlass' 2025-12-04T08:54:02.8714708Z Entering 'third_party/fbgemm' 2025-12-04T08:54:02.8735025Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:02.8753106Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:02.8781143Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:02.8800215Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:02.8833626Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:02.8858355Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:02.8875867Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:02.8895900Z Entering 'third_party/flash-attention' 2025-12-04T08:54:02.8914444Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:02.8935233Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:02.8957249Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:02.8977065Z Entering 'third_party/fmt' 2025-12-04T08:54:02.8996227Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:02.9016361Z Entering 'third_party/gloo' 2025-12-04T08:54:02.9038006Z Entering 'third_party/googletest' 2025-12-04T08:54:02.9056892Z Entering 'third_party/ideep' 2025-12-04T08:54:02.9076725Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:02.9097865Z Entering 'third_party/ittapi' 2025-12-04T08:54:02.9119308Z Entering 'third_party/kineto' 2025-12-04T08:54:02.9138263Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:02.9158635Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:02.9177436Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:02.9200694Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:02.9228637Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:02.9258622Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:02.9280049Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:02.9297106Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:02.9317701Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:02.9341771Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:02.9362427Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:02.9379994Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:02.9400241Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:02.9422893Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:02.9451772Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:02.9475033Z Entering 'third_party/kleidiai' 2025-12-04T08:54:02.9498151Z Entering 'third_party/mimalloc' 2025-12-04T08:54:02.9522008Z Entering 'third_party/nlohmann' 2025-12-04T08:54:02.9543611Z Entering 'third_party/onnx' 2025-12-04T08:54:02.9568215Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:02.9595106Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:02.9615757Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:02.9643004Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:02.9668688Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:02.9689009Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:02.9706490Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:02.9729933Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:02.9747682Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:02.9765167Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:02.9784726Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:02.9805846Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:02.9836860Z Entering 'third_party/pocketfft' 2025-12-04T08:54:02.9858865Z Entering 'third_party/protobuf' 2025-12-04T08:54:02.9879864Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:02.9899185Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:02.9919233Z Entering 'third_party/psimd' 2025-12-04T08:54:02.9937349Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:02.9954887Z Entering 'third_party/pybind11' 2025-12-04T08:54:02.9973419Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:02.9993013Z Entering 'third_party/sleef' 2025-12-04T08:54:03.0012373Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:03.0032764Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:03.0058457Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:03.0076382Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:03.0101259Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:03.0123822Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:03.0154645Z ##[endgroup] 2025-12-04T08:54:03.0300222Z [command]/usr/bin/git log -1 --format=%H 2025-12-04T08:54:03.0418366Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:54:03.0536853Z ##[group]Run actions/checkout@v4 2025-12-04T08:54:03.0536985Z with: 2025-12-04T08:54:03.0537099Z ref: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:54:03.0537227Z fetch-depth: 0 2025-12-04T08:54:03.0537325Z submodules: recursive 2025-12-04T08:54:03.0537424Z show-progress: false 2025-12-04T08:54:03.0537537Z repository: pytorch/pytorch 2025-12-04T08:54:03.0537692Z token: *** 2025-12-04T08:54:03.0537778Z ssh-strict: true 2025-12-04T08:54:03.0537869Z ssh-user: git 2025-12-04T08:54:03.0537963Z persist-credentials: true 2025-12-04T08:54:03.0538067Z clean: true 2025-12-04T08:54:03.0538162Z sparse-checkout-cone-mode: true 2025-12-04T08:54:03.0538290Z fetch-tags: false 2025-12-04T08:54:03.0538381Z lfs: false 2025-12-04T08:54:03.0538470Z set-safe-directory: true 2025-12-04T08:54:03.0538569Z env: 2025-12-04T08:54:03.0538657Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:03.0538754Z ##[endgroup] 2025-12-04T08:54:03.0988683Z Syncing repository: pytorch/pytorch 2025-12-04T08:54:03.0988974Z ##[group]Getting Git version info 2025-12-04T08:54:03.0989174Z Working directory is '/home/runner/_work/pytorch/pytorch' 2025-12-04T08:54:03.1001918Z [command]/usr/bin/git version 2025-12-04T08:54:03.1027004Z git version 2.52.0 2025-12-04T08:54:03.1047683Z ##[endgroup] 2025-12-04T08:54:03.1053417Z Copying '/home/runner/.gitconfig' to '/home/runner/_work/_temp/52dde6d9-2af2-4aaf-a9fd-ea1accac91fd/.gitconfig' 2025-12-04T08:54:03.1059265Z Temporarily overriding HOME='/home/runner/_work/_temp/52dde6d9-2af2-4aaf-a9fd-ea1accac91fd' before making global git config changes 2025-12-04T08:54:03.1059580Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T08:54:03.1062044Z [command]/usr/bin/git config --global --add safe.directory /home/runner/_work/pytorch/pytorch 2025-12-04T08:54:03.1084497Z [command]/usr/bin/git config --local --get remote.origin.url 2025-12-04T08:54:03.1099807Z https://github.com/pytorch/pytorch 2025-12-04T08:54:03.1112315Z ##[group]Removing previously created refs, to avoid conflicts 2025-12-04T08:54:03.1115547Z [command]/usr/bin/git rev-parse --symbolic-full-name --verify --quiet HEAD 2025-12-04T08:54:03.1131011Z HEAD 2025-12-04T08:54:03.1164854Z ##[endgroup] 2025-12-04T08:54:03.1167213Z [command]/usr/bin/git submodule status 2025-12-04T08:54:03.1372951Z 7e1e1fe3858c63c251c637ae41a20de425dde96f android/libs/fbjni (v0.1.0-12-g7e1e1fe) 2025-12-04T08:54:03.1424994Z 4dfe081cf6bcd15db339cf2680b9281b8451eeb3 third_party/FP16 (4dfe081) 2025-12-04T08:54:03.1478830Z b408327ac2a15ec3e43352421954f5b1967701d1 third_party/FXdiv (b408327) 2025-12-04T08:54:03.1532538Z c07e3a0400713d546e0dea2d5466dd22ea389c73 third_party/NNPACK (c07e3a0) 2025-12-04T08:54:03.1565917Z 3ebbc93ded7285963bff932c678fa367eb393ba6 third_party/NVTX (v3.1.0-313-g3ebbc93) 2025-12-04T08:54:03.1618134Z 1d8f600fd424278486eade7ed3e877c99f0846b1 third_party/VulkanMemoryAllocator (v2.1.0-982-g1d8f600) 2025-12-04T08:54:03.1916963Z 51a0103656eff6fc9bfd39a4597923c4b542c883 third_party/XNNPACK (remotes/origin/ds/ndk-1243-g51a0103656) 2025-12-04T08:54:03.1943714Z 01aae101b9e5e94d6c16a9514c9fb8df99c93150 third_party/aiter (v0.1.1-92-g01aae101) 2025-12-04T08:54:03.1962550Z 299e5928955cc62af9968370293b916f5130916f third_party/benchmark (v1.9.3) 2025-12-04T08:54:03.2016013Z 7fe50dc3da2069d6645d9deb8c017a876472a977 third_party/composable_kernel (rocm-6.4.3-459-g7fe50dc3d) 2025-12-04T08:54:03.2095439Z 89c932f313c6437c38f2982869beacc89c2f2246 third_party/cpp-httplib (v0.26.0) 2025-12-04T08:54:03.2169665Z f858c30bcb16f8effd5ff46996f0514539e17abc third_party/cpuinfo (f858c30) 2025-12-04T08:54:03.2199147Z 0b1577c8c83401237d601d0d0db5210506705396 third_party/cudnn_frontend (v0.5-61-g0b1577c) 2025-12-04T08:54:03.2269339Z f88806b1e31dfa579842638740216dd41fc6c588 third_party/cutlass (v4.3.1) 2025-12-04T08:54:03.2296822Z c0b988d39a9e47c794d699f29930ed4d7c7e13a4 third_party/fbgemm (v1.4.0-rc1-2-gc0b988d39) 2025-12-04T08:54:03.2347720Z 979702c87a8713a8e0a5e9fee122b90d2ef13be5 third_party/flash-attention (v2.7.4) 2025-12-04T08:54:03.2376563Z a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757 third_party/flatbuffers (v24.12.23) 2025-12-04T08:54:03.2639063Z 407c905e45ad75fc29bf0f9bb7c5c2fd3475976f third_party/fmt (12.1.0) 2025-12-04T08:54:03.2696228Z 3fb5c176c17c765a3492cd2f0321b0dab712f350 third_party/gemmlowp/gemmlowp (remotes/origin/revert-87-master-135-g3fb5c17) 2025-12-04T08:54:03.2787949Z 54cbae0d3a67fa890b4c3d9ee162b7860315e341 third_party/gloo (remotes/origin/gh/c-p-i-o/1/base-37-g54cbae0) 2025-12-04T08:54:03.2944256Z 52eb8108c5bdec04579160ae17225d66034bd723 third_party/googletest (release-1.8.0-3544-g52eb8108) 2025-12-04T08:54:03.2994906Z 719d8e6cd7f7a0e01b155657526d693acf97c2b3 third_party/ideep (pytorch-rls-v3.7.1) 2025-12-04T08:54:03.3047766Z dec1d23ca65ab069d225dfe40dea14f455170959 third_party/ittapi (v3.25.5) 2025-12-04T08:54:03.3184313Z 31f85df8fbd89c188f14ef10f1ec65379786b943 third_party/kineto (heads/main) 2025-12-04T08:54:03.3204898Z d7770c89632329a9914ef1a90289917597639cbe third_party/kleidiai (v1.15.0) 2025-12-04T08:54:03.3223135Z fbd8b99c2b828428947d70fdc046bb55609be93e third_party/mimalloc (v2.2.4) 2025-12-04T08:54:03.3246217Z 55f93686c01528224f448c19128836e7df245f72 third_party/nlohmann (v3.12.0) 2025-12-04T08:54:03.3455929Z e709452ef2bbc1d113faf678c24e6d3467696e83 third_party/onnx (v1.18.0) 2025-12-04T08:54:03.3478945Z a799f4aed9c94b765dcdaabaeab7d5e7e2310878 third_party/opentelemetry-cpp (v1.14.2) 2025-12-04T08:54:03.3505294Z 0fa0ef591e38c2758e3184c6c23e497b9f732ffa third_party/pocketfft (release_for_eigen-40-g0fa0ef5) 2025-12-04T08:54:03.3716504Z d1eca4e4b421cd2997495c4b4e65cea6be4e9b8a third_party/protobuf (v3.7.0-rc.2-1279-gd1eca4e4b) 2025-12-04T08:54:03.3765467Z 072586a71b55b7f8c584153d223e95687148a900 third_party/psimd (heads/master) 2025-12-04T08:54:03.3807445Z 4fe0e1e183925bf8cfa6aae24237e724a96479b8 third_party/pthreadpool (0.1-144-g4fe0e1e) 2025-12-04T08:54:03.3838232Z f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8 third_party/pybind11 (v3.0.1) 2025-12-04T08:54:03.3908911Z f45429b087dd7d5bc78bb40dc7cf06425c252d67 third_party/python-peachpy (remotes/origin/pre-generated) 2025-12-04T08:54:03.3971269Z 5a1d179df9cf652951b59010a2d2075372d67f68 third_party/sleef (3.8) 2025-12-04T08:54:03.4027514Z 2b4cd91092d335a697416b2a3cb398283246849d third_party/tensorpipe (heads/main) 2025-12-04T08:54:03.4038668Z ##[group]Cleaning the repository 2025-12-04T08:54:03.4044465Z [command]/usr/bin/git clean -ffdx 2025-12-04T08:54:03.4172447Z [command]/usr/bin/git reset --hard HEAD 2025-12-04T08:54:03.4939707Z HEAD is now at ffd9b0fb4355 Resolve collective autotuning test failure on arm (#168919) 2025-12-04T08:54:03.5000302Z ##[endgroup] 2025-12-04T08:54:03.5003632Z ##[group]Disabling automatic garbage collection 2025-12-04T08:54:03.5010269Z [command]/usr/bin/git config --local gc.auto 0 2025-12-04T08:54:03.5034335Z ##[endgroup] 2025-12-04T08:54:03.5034624Z ##[group]Setting up auth 2025-12-04T08:54:03.5038765Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T08:54:03.5068159Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T08:54:03.5284049Z Entering 'android/libs/fbjni' 2025-12-04T08:54:03.5317262Z Entering 'third_party/FP16' 2025-12-04T08:54:03.5344775Z Entering 'third_party/FXdiv' 2025-12-04T08:54:03.5378704Z Entering 'third_party/NNPACK' 2025-12-04T08:54:03.5408730Z Entering 'third_party/NVTX' 2025-12-04T08:54:03.5441450Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:03.5472196Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:03.5512069Z Entering 'third_party/aiter' 2025-12-04T08:54:03.5539133Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:03.5571317Z Entering 'third_party/benchmark' 2025-12-04T08:54:03.5599373Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:03.5640441Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:03.5671865Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:03.5702747Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:03.5729523Z Entering 'third_party/cutlass' 2025-12-04T08:54:03.5764221Z Entering 'third_party/fbgemm' 2025-12-04T08:54:03.5796200Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:03.5821621Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:03.5857341Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:03.5881622Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:03.5916109Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:03.5943887Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:03.5969901Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:03.6010097Z Entering 'third_party/flash-attention' 2025-12-04T08:54:03.6042537Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:03.6076959Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:03.6106064Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:03.6138053Z Entering 'third_party/fmt' 2025-12-04T08:54:03.6166056Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:03.6190247Z Entering 'third_party/gloo' 2025-12-04T08:54:03.6215776Z Entering 'third_party/googletest' 2025-12-04T08:54:03.6242125Z Entering 'third_party/ideep' 2025-12-04T08:54:03.6270137Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:03.6294087Z Entering 'third_party/ittapi' 2025-12-04T08:54:03.6321285Z Entering 'third_party/kineto' 2025-12-04T08:54:03.6351315Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:03.6376507Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:03.6399046Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:03.6429635Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:03.6460366Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:03.6492274Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:03.6521344Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:03.6546977Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:03.6576424Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:03.6600628Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:03.6624589Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:03.6650019Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:03.6672873Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:03.6707647Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:03.6730204Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:03.6753853Z Entering 'third_party/kleidiai' 2025-12-04T08:54:03.6780895Z Entering 'third_party/mimalloc' 2025-12-04T08:54:03.6808963Z Entering 'third_party/nlohmann' 2025-12-04T08:54:03.6844232Z Entering 'third_party/onnx' 2025-12-04T08:54:03.6875453Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:03.6906936Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:03.6937110Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:03.6959722Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:03.6987394Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:03.7015551Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:03.7039985Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:03.7073897Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:03.7099973Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:03.7125851Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:03.7149722Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:03.7175074Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:03.7217445Z Entering 'third_party/pocketfft' 2025-12-04T08:54:03.7253584Z Entering 'third_party/protobuf' 2025-12-04T08:54:03.7281525Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:03.7305692Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:03.7335814Z Entering 'third_party/psimd' 2025-12-04T08:54:03.7371331Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:03.7395604Z Entering 'third_party/pybind11' 2025-12-04T08:54:03.7432411Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:03.7458815Z Entering 'third_party/sleef' 2025-12-04T08:54:03.7487921Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:03.7513021Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:03.7539071Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:03.7567114Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:03.7592391Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:03.7619844Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:03.7667198Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T08:54:03.7690411Z http.https://github.com/.extraheader 2025-12-04T08:54:03.7702567Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-12-04T08:54:03.7724988Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T08:54:03.7900019Z Entering 'android/libs/fbjni' 2025-12-04T08:54:03.7919916Z http.https://github.com/.extraheader 2025-12-04T08:54:03.7936620Z Entering 'third_party/FP16' 2025-12-04T08:54:03.7951867Z http.https://github.com/.extraheader 2025-12-04T08:54:03.7969346Z Entering 'third_party/FXdiv' 2025-12-04T08:54:03.7983216Z http.https://github.com/.extraheader 2025-12-04T08:54:03.7999097Z Entering 'third_party/NNPACK' 2025-12-04T08:54:03.8011306Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8033790Z Entering 'third_party/NVTX' 2025-12-04T08:54:03.8048679Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8067912Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:03.8080795Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8100195Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:03.8116955Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8144771Z Entering 'third_party/aiter' 2025-12-04T08:54:03.8163448Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8178991Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:03.8194243Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8221068Z Entering 'third_party/benchmark' 2025-12-04T08:54:03.8234178Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8248171Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:03.8259077Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8282215Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:03.8294013Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8311028Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:03.8323422Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8340684Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:03.8359499Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8377434Z Entering 'third_party/cutlass' 2025-12-04T08:54:03.8390275Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8410741Z Entering 'third_party/fbgemm' 2025-12-04T08:54:03.8422973Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8439142Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:03.8452806Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8471588Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:03.8488488Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8507503Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:03.8527725Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8550363Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:03.8562534Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8586375Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:03.8598559Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8616617Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:03.8629514Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8644554Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:03.8656540Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8675509Z Entering 'third_party/flash-attention' 2025-12-04T08:54:03.8688358Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8705890Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:03.8723424Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8752213Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:03.8766013Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8794652Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:03.8809644Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8827253Z Entering 'third_party/fmt' 2025-12-04T08:54:03.8839260Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8857084Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:03.8868785Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8886335Z Entering 'third_party/gloo' 2025-12-04T08:54:03.8899772Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8918140Z Entering 'third_party/googletest' 2025-12-04T08:54:03.8931478Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8948911Z Entering 'third_party/ideep' 2025-12-04T08:54:03.8961293Z http.https://github.com/.extraheader 2025-12-04T08:54:03.8975057Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:03.8986256Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9005762Z Entering 'third_party/ittapi' 2025-12-04T08:54:03.9020209Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9041780Z Entering 'third_party/kineto' 2025-12-04T08:54:03.9055395Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9072090Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:03.9085195Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9102257Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:03.9114633Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9133639Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:03.9144640Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9161267Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:03.9173377Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9191785Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:03.9224358Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9244929Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:03.9266132Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9285672Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:03.9299134Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9321130Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:03.9335101Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9357137Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:03.9372426Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9391121Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:03.9405181Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9426072Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:03.9439579Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9457326Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:03.9469605Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9485602Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:03.9497755Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9525706Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:03.9537992Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9553730Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:03.9566326Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9583007Z Entering 'third_party/kleidiai' 2025-12-04T08:54:03.9595406Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9609827Z Entering 'third_party/mimalloc' 2025-12-04T08:54:03.9621327Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9637938Z Entering 'third_party/nlohmann' 2025-12-04T08:54:03.9651940Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9668098Z Entering 'third_party/onnx' 2025-12-04T08:54:03.9679968Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9701804Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:03.9717392Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9737699Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:03.9751380Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9769558Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:03.9781477Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9799236Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:03.9819630Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9837609Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:03.9853211Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9868700Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:03.9879946Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9895758Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:03.9906895Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9923281Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:03.9935447Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9950670Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:03.9967141Z http.https://github.com/.extraheader 2025-12-04T08:54:03.9982062Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:03.9996969Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0016545Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:04.0030119Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0050146Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:04.0064785Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0092553Z Entering 'third_party/pocketfft' 2025-12-04T08:54:04.0105555Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0126036Z Entering 'third_party/protobuf' 2025-12-04T08:54:04.0141779Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0160575Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:04.0173410Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0195183Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:04.0207413Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0227466Z Entering 'third_party/psimd' 2025-12-04T08:54:04.0238796Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0254340Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:04.0269045Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0288795Z Entering 'third_party/pybind11' 2025-12-04T08:54:04.0304198Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0321001Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:04.0333473Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0356170Z Entering 'third_party/sleef' 2025-12-04T08:54:04.0369706Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0385199Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:04.0398504Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0418643Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:04.0430882Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0453313Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:04.0465828Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0487006Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:04.0498803Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0514495Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:04.0527287Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0544311Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:04.0556799Z http.https://github.com/.extraheader 2025-12-04T08:54:04.0593453Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.0616106Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T08:54:04.0776091Z Entering 'android/libs/fbjni' 2025-12-04T08:54:04.0785665Z file:/home/runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T08:54:04.0794733Z Entering 'third_party/FP16' 2025-12-04T08:54:04.0804469Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T08:54:04.0817024Z Entering 'third_party/FXdiv' 2025-12-04T08:54:04.0829928Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T08:54:04.0842466Z Entering 'third_party/NNPACK' 2025-12-04T08:54:04.0852560Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T08:54:04.0861517Z Entering 'third_party/NVTX' 2025-12-04T08:54:04.0874670Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T08:54:04.0884707Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:04.0894967Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T08:54:04.0903139Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:04.0912060Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T08:54:04.0925270Z Entering 'third_party/aiter' 2025-12-04T08:54:04.0934537Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T08:54:04.0943297Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:04.0951721Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T08:54:04.0964403Z Entering 'third_party/benchmark' 2025-12-04T08:54:04.0976514Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:54:04.0985482Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:04.0995686Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T08:54:04.1006553Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:04.1016479Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T08:54:04.1025131Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:04.1034282Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T08:54:04.1042673Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:04.1051959Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T08:54:04.1060612Z Entering 'third_party/cutlass' 2025-12-04T08:54:04.1069985Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T08:54:04.1083238Z Entering 'third_party/fbgemm' 2025-12-04T08:54:04.1093132Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T08:54:04.1102287Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:04.1128480Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T08:54:04.1138393Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:04.1148499Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T08:54:04.1159148Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:04.1169897Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T08:54:04.1185472Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:04.1197506Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T08:54:04.1209808Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:04.1224144Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T08:54:04.1233261Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:04.1242366Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T08:54:04.1250509Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:04.1259815Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T08:54:04.1271204Z Entering 'third_party/flash-attention' 2025-12-04T08:54:04.1280848Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T08:54:04.1288864Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:04.1298641Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T08:54:04.1312313Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:04.1321452Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T08:54:04.1334163Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:04.1349644Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T08:54:04.1359717Z Entering 'third_party/fmt' 2025-12-04T08:54:04.1370435Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T08:54:04.1380597Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:04.1390289Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T08:54:04.1400556Z Entering 'third_party/gloo' 2025-12-04T08:54:04.1410094Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T08:54:04.1420555Z Entering 'third_party/googletest' 2025-12-04T08:54:04.1434645Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:04.1444169Z Entering 'third_party/ideep' 2025-12-04T08:54:04.1455511Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T08:54:04.1463942Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:04.1483430Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T08:54:04.1496726Z Entering 'third_party/ittapi' 2025-12-04T08:54:04.1506021Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T08:54:04.1515332Z Entering 'third_party/kineto' 2025-12-04T08:54:04.1525016Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T08:54:04.1533183Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:04.1542662Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T08:54:04.1552647Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:04.1570443Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T08:54:04.1579760Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:04.1589632Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T08:54:04.1603108Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:04.1612673Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T08:54:04.1621208Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:04.1631924Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T08:54:04.1640379Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:04.1659651Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T08:54:04.1670532Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:04.1680054Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T08:54:04.1688000Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:04.1702812Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:04.1711075Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:04.1720990Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T08:54:04.1729424Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:04.1738251Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T08:54:04.1746545Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:04.1755716Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T08:54:04.1763841Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:04.1772659Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T08:54:04.1781984Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:04.1790818Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T08:54:04.1802221Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:04.1811120Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T08:54:04.1825562Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:04.1835902Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T08:54:04.1846867Z Entering 'third_party/kleidiai' 2025-12-04T08:54:04.1857450Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T08:54:04.1869002Z Entering 'third_party/mimalloc' 2025-12-04T08:54:04.1882883Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T08:54:04.1891522Z Entering 'third_party/nlohmann' 2025-12-04T08:54:04.1902163Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T08:54:04.1911351Z Entering 'third_party/onnx' 2025-12-04T08:54:04.1920787Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T08:54:04.1943338Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:04.1959908Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:54:04.1973821Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:04.1984468Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T08:54:04.1993513Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:04.2013129Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:54:04.2027888Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:04.2037222Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:04.2045586Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:04.2055449Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T08:54:04.2063504Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:04.2079091Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T08:54:04.2088223Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:04.2103054Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T08:54:04.2111693Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:04.2120766Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T08:54:04.2132110Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:04.2142190Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T08:54:04.2152193Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:04.2164323Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T08:54:04.2174584Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:04.2188819Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T08:54:04.2198137Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:04.2207634Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T08:54:04.2223401Z Entering 'third_party/pocketfft' 2025-12-04T08:54:04.2233110Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T08:54:04.2241386Z Entering 'third_party/protobuf' 2025-12-04T08:54:04.2251132Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T08:54:04.2260995Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:04.2270423Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:54:04.2278676Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:04.2289455Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:04.2299311Z Entering 'third_party/psimd' 2025-12-04T08:54:04.2308927Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T08:54:04.2317791Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:04.2326882Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T08:54:04.2337461Z Entering 'third_party/pybind11' 2025-12-04T08:54:04.2347885Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:54:04.2360885Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:04.2371181Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T08:54:04.2385509Z Entering 'third_party/sleef' 2025-12-04T08:54:04.2396995Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T08:54:04.2404823Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:04.2414717Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T08:54:04.2425259Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:04.2440553Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:04.2457136Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:04.2466786Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T08:54:04.2474862Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:04.2483871Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T08:54:04.2492068Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:04.2501036Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:54:04.2509013Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:04.2530195Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T08:54:04.2559071Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2584762Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2607057Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2628658Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2647911Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2668819Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2688712Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2705633Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2722881Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2738099Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2759821Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2776026Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2796622Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2816210Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2831561Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2846697Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2861421Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2877341Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2893353Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2910233Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2925158Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2940449Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2955135Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2970134Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.2986348Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3000575Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3015123Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3029125Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3043840Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3062732Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3077771Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3091353Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3106274Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3121985Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3138435Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3158322Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3174066Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3195028Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3218091Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3233962Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3252651Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3268856Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3291197Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3313047Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3338690Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3372068Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3395944Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3421273Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3439325Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3462042Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3478225Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3496133Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3515561Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3536921Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3561067Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3580579Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3605680Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3627421Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3646585Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3662882Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3685302Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3702504Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3720995Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3739582Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3756126Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3773871Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3797551Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3816336Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3844781Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3864103Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3881339Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3900267Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3916308Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3934646Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3953261Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3980663Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.3997253Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.4020392Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.4038844Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.4060205Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.4078044Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T08:54:04.4097973Z [command]/usr/bin/git config --local http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T08:54:04.4130711Z ##[endgroup] 2025-12-04T08:54:04.4130933Z ##[group]Fetching the repository 2025-12-04T08:54:04.4140562Z [command]/usr/bin/git -c protocol.version=2 fetch --prune --no-recurse-submodules origin +refs/heads/*:refs/remotes/origin/* +refs/tags/*:refs/tags/* 2025-12-04T08:54:05.8419767Z [command]/usr/bin/git rev-parse --verify --quiet ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32^{object} 2025-12-04T08:54:05.8879457Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:54:05.8883568Z ##[endgroup] 2025-12-04T08:54:05.8883909Z ##[group]Determining the checkout info 2025-12-04T08:54:05.8885279Z ##[endgroup] 2025-12-04T08:54:05.8890182Z [command]/usr/bin/git sparse-checkout disable 2025-12-04T08:54:05.8974588Z [command]/usr/bin/git config --local --unset-all extensions.worktreeConfig 2025-12-04T08:54:05.8989140Z ##[group]Checking out the ref 2025-12-04T08:54:05.8990386Z [command]/usr/bin/git checkout --progress --force ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:54:05.9252417Z HEAD is now at ffd9b0fb4355 Resolve collective autotuning test failure on arm (#168919) 2025-12-04T08:54:05.9257003Z ##[endgroup] 2025-12-04T08:54:05.9257386Z ##[group]Setting up auth for fetching submodules 2025-12-04T08:54:05.9260372Z [command]/usr/bin/git config --global http.https://github.com/.extraheader AUTHORIZATION: basic *** 2025-12-04T08:54:05.9287281Z [command]/usr/bin/git config --global --unset-all url.https://github.com/.insteadOf 2025-12-04T08:54:05.9308687Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf git@github.com: 2025-12-04T08:54:05.9325108Z [command]/usr/bin/git config --global --add url.https://github.com/.insteadOf org-21003710@github.com: 2025-12-04T08:54:05.9344897Z ##[endgroup] 2025-12-04T08:54:05.9345156Z ##[group]Fetching submodules 2025-12-04T08:54:05.9346117Z [command]/usr/bin/git submodule sync --recursive 2025-12-04T08:54:05.9531105Z Synchronizing submodule url for 'android/libs/fbjni' 2025-12-04T08:54:05.9541873Z Synchronizing submodule url for 'third_party/FP16' 2025-12-04T08:54:05.9556709Z Synchronizing submodule url for 'third_party/FXdiv' 2025-12-04T08:54:05.9568584Z Synchronizing submodule url for 'third_party/NNPACK' 2025-12-04T08:54:05.9578460Z Synchronizing submodule url for 'third_party/NVTX' 2025-12-04T08:54:05.9595970Z Synchronizing submodule url for 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:05.9607358Z Synchronizing submodule url for 'third_party/XNNPACK' 2025-12-04T08:54:05.9623813Z Synchronizing submodule url for 'third_party/aiter' 2025-12-04T08:54:05.9644310Z Synchronizing submodule url for 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:05.9660844Z Synchronizing submodule url for 'third_party/benchmark' 2025-12-04T08:54:05.9671085Z Synchronizing submodule url for 'third_party/composable_kernel' 2025-12-04T08:54:05.9684969Z Synchronizing submodule url for 'third_party/cpp-httplib' 2025-12-04T08:54:05.9701197Z Synchronizing submodule url for 'third_party/cpuinfo' 2025-12-04T08:54:05.9711565Z Synchronizing submodule url for 'third_party/cudnn_frontend' 2025-12-04T08:54:05.9728265Z Synchronizing submodule url for 'third_party/cutlass' 2025-12-04T08:54:05.9742316Z Synchronizing submodule url for 'third_party/fbgemm' 2025-12-04T08:54:05.9758855Z Synchronizing submodule url for 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:05.9775866Z Synchronizing submodule url for 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:05.9791136Z Synchronizing submodule url for 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:05.9801123Z Synchronizing submodule url for 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:05.9818097Z Synchronizing submodule url for 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:05.9828844Z Synchronizing submodule url for 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:05.9841383Z Synchronizing submodule url for 'third_party/fbgemm/external/json' 2025-12-04T08:54:05.9856542Z Synchronizing submodule url for 'third_party/flash-attention' 2025-12-04T08:54:05.9877322Z Synchronizing submodule url for 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:05.9889199Z Synchronizing submodule url for 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:05.9905712Z Synchronizing submodule url for 'third_party/flatbuffers' 2025-12-04T08:54:05.9919430Z Synchronizing submodule url for 'third_party/fmt' 2025-12-04T08:54:05.9929834Z Synchronizing submodule url for 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:05.9940672Z Synchronizing submodule url for 'third_party/gloo' 2025-12-04T08:54:05.9951239Z Synchronizing submodule url for 'third_party/googletest' 2025-12-04T08:54:05.9961627Z Synchronizing submodule url for 'third_party/ideep' 2025-12-04T08:54:05.9973217Z Synchronizing submodule url for 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:05.9987599Z Synchronizing submodule url for 'third_party/ittapi' 2025-12-04T08:54:05.9998232Z Synchronizing submodule url for 'third_party/kineto' 2025-12-04T08:54:06.0010735Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:06.0020114Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:06.0031135Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:06.0041463Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:06.0051855Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:06.0062892Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:06.0076214Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:06.0086458Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:06.0103502Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:06.0114260Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:06.0124330Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:06.0135240Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:06.0146436Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:06.0165909Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:06.0185934Z Synchronizing submodule url for 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:06.0199144Z Synchronizing submodule url for 'third_party/kleidiai' 2025-12-04T08:54:06.0209877Z Synchronizing submodule url for 'third_party/mimalloc' 2025-12-04T08:54:06.0221360Z Synchronizing submodule url for 'third_party/nlohmann' 2025-12-04T08:54:06.0239179Z Synchronizing submodule url for 'third_party/onnx' 2025-12-04T08:54:06.0257878Z Synchronizing submodule url for 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:06.0270110Z Synchronizing submodule url for 'third_party/opentelemetry-cpp' 2025-12-04T08:54:06.0282390Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:06.0291380Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:06.0301534Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:06.0311841Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:06.0323526Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:06.0336380Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:06.0346282Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:06.0357195Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:06.0374384Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:06.0385904Z Synchronizing submodule url for 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:06.0405594Z Synchronizing submodule url for 'third_party/pocketfft' 2025-12-04T08:54:06.0416012Z Synchronizing submodule url for 'third_party/protobuf' 2025-12-04T08:54:06.0428710Z Synchronizing submodule url for 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:06.0442287Z Synchronizing submodule url for 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:06.0454724Z Synchronizing submodule url for 'third_party/psimd' 2025-12-04T08:54:06.0464779Z Synchronizing submodule url for 'third_party/pthreadpool' 2025-12-04T08:54:06.0475094Z Synchronizing submodule url for 'third_party/pybind11' 2025-12-04T08:54:06.0485109Z Synchronizing submodule url for 'third_party/python-peachpy' 2025-12-04T08:54:06.0494112Z Synchronizing submodule url for 'third_party/sleef' 2025-12-04T08:54:06.0503370Z Synchronizing submodule url for 'third_party/tensorpipe' 2025-12-04T08:54:06.0522201Z Synchronizing submodule url for 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:06.0533330Z Synchronizing submodule url for 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:06.0543946Z Synchronizing submodule url for 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:06.0554324Z Synchronizing submodule url for 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:06.0572304Z Synchronizing submodule url for 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:06.0598125Z [command]/usr/bin/git -c protocol.version=2 submodule update --init --force --recursive 2025-12-04T08:54:06.0803415Z Submodule path 'android/libs/fbjni': checked out '7e1e1fe3858c63c251c637ae41a20de425dde96f' 2025-12-04T08:54:06.0862604Z Submodule path 'third_party/FP16': checked out '4dfe081cf6bcd15db339cf2680b9281b8451eeb3' 2025-12-04T08:54:06.0910439Z Submodule path 'third_party/FXdiv': checked out 'b408327ac2a15ec3e43352421954f5b1967701d1' 2025-12-04T08:54:06.0959971Z Submodule path 'third_party/NNPACK': checked out 'c07e3a0400713d546e0dea2d5466dd22ea389c73' 2025-12-04T08:54:06.1025779Z Submodule path 'third_party/NVTX': checked out '3ebbc93ded7285963bff932c678fa367eb393ba6' 2025-12-04T08:54:06.1087505Z Submodule path 'third_party/VulkanMemoryAllocator': checked out '1d8f600fd424278486eade7ed3e877c99f0846b1' 2025-12-04T08:54:06.1226115Z Submodule path 'third_party/XNNPACK': checked out '51a0103656eff6fc9bfd39a4597923c4b542c883' 2025-12-04T08:54:06.1360493Z Submodule path 'third_party/aiter': checked out '01aae101b9e5e94d6c16a9514c9fb8df99c93150' 2025-12-04T08:54:06.1538094Z Submodule path 'third_party/aiter/3rdparty/composable_kernel': checked out 'cffe8fa2a442ac8e80dd236a1a5d24fe3d7e0cbf' 2025-12-04T08:54:06.1605090Z Submodule path 'third_party/benchmark': checked out '299e5928955cc62af9968370293b916f5130916f' 2025-12-04T08:54:06.1773631Z Submodule path 'third_party/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T08:54:06.1839732Z Submodule path 'third_party/cpp-httplib': checked out '89c932f313c6437c38f2982869beacc89c2f2246' 2025-12-04T08:54:06.1897771Z Submodule path 'third_party/cpuinfo': checked out 'f858c30bcb16f8effd5ff46996f0514539e17abc' 2025-12-04T08:54:06.1967468Z Submodule path 'third_party/cudnn_frontend': checked out '0b1577c8c83401237d601d0d0db5210506705396' 2025-12-04T08:54:06.2070589Z Submodule path 'third_party/cutlass': checked out 'f88806b1e31dfa579842638740216dd41fc6c588' 2025-12-04T08:54:06.2185676Z Submodule path 'third_party/fbgemm': checked out 'c0b988d39a9e47c794d699f29930ed4d7c7e13a4' 2025-12-04T08:54:06.2233660Z Submodule path 'third_party/fbgemm/external/asmjit': checked out 'a3199e8857792cd10b7589ff5d58343d2c9008ea' 2025-12-04T08:54:06.2406525Z Submodule path 'third_party/fbgemm/external/composable_kernel': checked out '7fe50dc3da2069d6645d9deb8c017a876472a977' 2025-12-04T08:54:06.2472032Z Submodule path 'third_party/fbgemm/external/cpuinfo': checked out '6543fec09b2f04ac4a666882998b534afc9c1349' 2025-12-04T08:54:06.2581609Z Submodule path 'third_party/fbgemm/external/cutlass': checked out '98125ce499b0fdf7ffbe0e3052f5b8709f4840f8' 2025-12-04T08:54:06.2648268Z Submodule path 'third_party/fbgemm/external/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:54:06.2700832Z Submodule path 'third_party/fbgemm/external/hipify_torch': checked out '63b6a7b541fa7f08f8475ca7d74054db36ff2691' 2025-12-04T08:54:06.2776582Z Submodule path 'third_party/fbgemm/external/json': checked out '9cca280a4d0ccf0c08f47a99aa71d1b0e52f8d03' 2025-12-04T08:54:06.2851681Z Submodule path 'third_party/flash-attention': checked out '979702c87a8713a8e0a5e9fee122b90d2ef13be5' 2025-12-04T08:54:06.3018126Z Submodule path 'third_party/flash-attention/csrc/composable_kernel': checked out '888317e698e9803c62bd38568abc9e05d7709f33' 2025-12-04T08:54:06.3122544Z Submodule path 'third_party/flash-attention/csrc/cutlass': checked out 'c506e16788cb08416a4a57e11a9067beeee29420' 2025-12-04T08:54:06.3216637Z Submodule path 'third_party/flatbuffers': checked out 'a2cd1ea3b6d3fee220106b5fed3f7ce8da9eb757' 2025-12-04T08:54:06.3280996Z Submodule path 'third_party/fmt': checked out '407c905e45ad75fc29bf0f9bb7c5c2fd3475976f' 2025-12-04T08:54:06.3330350Z Submodule path 'third_party/gemmlowp/gemmlowp': checked out '3fb5c176c17c765a3492cd2f0321b0dab712f350' 2025-12-04T08:54:06.3382555Z Submodule path 'third_party/gloo': checked out '54cbae0d3a67fa890b4c3d9ee162b7860315e341' 2025-12-04T08:54:06.3436984Z Submodule path 'third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:54:06.3489871Z Submodule path 'third_party/ideep': checked out '719d8e6cd7f7a0e01b155657526d693acf97c2b3' 2025-12-04T08:54:06.3643842Z Submodule path 'third_party/ideep/mkl-dnn': checked out '8d263e693366ef8db40acc569cc7d8edf644556d' 2025-12-04T08:54:06.3700773Z Submodule path 'third_party/ittapi': checked out 'dec1d23ca65ab069d225dfe40dea14f455170959' 2025-12-04T08:54:06.3778471Z Submodule path 'third_party/kineto': checked out '31f85df8fbd89c188f14ef10f1ec65379786b943' 2025-12-04T08:54:06.3873151Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog': checked out 'd2ffe0a4e3acace628db49974246b66fc3e85fb1' 2025-12-04T08:54:06.3950306Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM': checked out 'ffde4e54bc7249a6039a5e6b45b395141e1217f9' 2025-12-04T08:54:06.4019146Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr': checked out '871ed52d350214a034f6ef8a3b8f51c5ce1bd400' 2025-12-04T08:54:06.4078182Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt': checked out 'cd4af11efc9c622896a3e4cb599fa28668ca3d05' 2025-12-04T08:54:06.4134482Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags': checked out 'e171aa2d15ed9eb17054558e0b3a6a413bb01067' 2025-12-04T08:54:06.4189839Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc': checked out '8411df715cf522606e3b1aca386ddfc0b63d34b4' 2025-12-04T08:54:06.4244587Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog': checked out 'b33e3bad4c46c8a6345525fd822af355e5ef9446' 2025-12-04T08:54:06.4298566Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:54:06.4384109Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/json': checked out '4f8fba14066156b73f1189a2b8bd568bde5284c5' 2025-12-04T08:54:06.4433809Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs': checked out 'f68a2fa8ea36c783bdd760371411fcb495aa3150' 2025-12-04T08:54:06.4488452Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp': checked out 'b1234816facfdda29845c46696a02998a4af115a' 2025-12-04T08:54:06.4572802Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'd7ba35bbb649209c66e582d5a0244ba988a15159' 2025-12-04T08:54:06.4635083Z Submodule path 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T08:54:06.4722838Z Submodule path 'third_party/kineto/libkineto/third_party/fmt': checked out '40626af88bd7df9a5fb80be7b25ac85b122d6c21' 2025-12-04T08:54:06.4784271Z Submodule path 'third_party/kineto/libkineto/third_party/googletest': checked out '52eb8108c5bdec04579160ae17225d66034bd723' 2025-12-04T08:54:06.4868911Z Submodule path 'third_party/kleidiai': checked out 'd7770c89632329a9914ef1a90289917597639cbe' 2025-12-04T08:54:06.4935849Z Submodule path 'third_party/mimalloc': checked out 'fbd8b99c2b828428947d70fdc046bb55609be93e' 2025-12-04T08:54:06.5025538Z Submodule path 'third_party/nlohmann': checked out '55f93686c01528224f448c19128836e7df245f72' 2025-12-04T08:54:06.5162607Z Submodule path 'third_party/onnx': checked out 'e709452ef2bbc1d113faf678c24e6d3467696e83' 2025-12-04T08:54:06.5247013Z Submodule path 'third_party/onnx/third_party/pybind11': checked out 'a2e59f0e7065404b44dfe92a28aca47ba1378dc4' 2025-12-04T08:54:06.5348182Z Submodule path 'third_party/opentelemetry-cpp': checked out 'a799f4aed9c94b765dcdaabaeab7d5e7e2310878' 2025-12-04T08:54:06.5417058Z Submodule path 'third_party/opentelemetry-cpp/third_party/benchmark': checked out 'd572f4777349d43653b21d6c2fc63020ab326db2' 2025-12-04T08:54:06.5477207Z Submodule path 'third_party/opentelemetry-cpp/third_party/googletest': checked out 'b796f7d44681514f58a683a3a71ff17c94edb0c1' 2025-12-04T08:54:06.5531827Z Submodule path 'third_party/opentelemetry-cpp/third_party/ms-gsl': checked out '6f4529395c5b7c2d661812257cd6780c67e54afa' 2025-12-04T08:54:06.5613011Z Submodule path 'third_party/opentelemetry-cpp/third_party/nlohmann-json': checked out 'bc889afb4c5bf1c0d8ee29ef35eaaf4c8bef8a5d' 2025-12-04T08:54:06.5684965Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto': checked out '4ca4f0335c63cda7ab31ea7ed70d6553aee14dce' 2025-12-04T08:54:06.5731509Z Submodule path 'third_party/opentelemetry-cpp/third_party/opentracing-cpp': checked out '06b57f48ded1fa3bdd3d4346f6ef29e40e08eaf5' 2025-12-04T08:54:06.5798275Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp': checked out 'c9ffcdda9086ffd9e1283ea7a0276d831f3c8a8d' 2025-12-04T08:54:06.5875786Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb': checked out 'eefb26f82b233268fc98577d265352720d477ba4' 2025-12-04T08:54:06.5929007Z Submodule path 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest': checked out 'e2239ee6043f73722e7aa812a459f54a28552929' 2025-12-04T08:54:06.6089389Z Submodule path 'third_party/opentelemetry-cpp/tools/vcpkg': checked out '8eb57355a4ffb410a2e94c07b4dca2dffbee8e50' 2025-12-04T08:54:06.6170392Z Submodule path 'third_party/pocketfft': checked out '0fa0ef591e38c2758e3184c6c23e497b9f732ffa' 2025-12-04T08:54:06.6329851Z Submodule path 'third_party/protobuf': checked out 'd1eca4e4b421cd2997495c4b4e65cea6be4e9b8a' 2025-12-04T08:54:06.6390244Z Submodule path 'third_party/protobuf/third_party/benchmark': checked out '5b7683f49e1e9223cf9927b24f6fd3d6bd82e3f8' 2025-12-04T08:54:06.6444417Z Submodule path 'third_party/protobuf/third_party/googletest': checked out '5ec7f0c4a113e2f18ac2c6cc7df51ad6afc24081' 2025-12-04T08:54:06.6493781Z Submodule path 'third_party/psimd': checked out '072586a71b55b7f8c584153d223e95687148a900' 2025-12-04T08:54:06.6548095Z Submodule path 'third_party/pthreadpool': checked out '4fe0e1e183925bf8cfa6aae24237e724a96479b8' 2025-12-04T08:54:06.6625609Z Submodule path 'third_party/pybind11': checked out 'f5fbe867d2d26e4a0a9177a51f6e568868ad3dc8' 2025-12-04T08:54:06.6677984Z Submodule path 'third_party/python-peachpy': checked out 'f45429b087dd7d5bc78bb40dc7cf06425c252d67' 2025-12-04T08:54:06.6731693Z Submodule path 'third_party/sleef': checked out '5a1d179df9cf652951b59010a2d2075372d67f68' 2025-12-04T08:54:06.6792235Z Submodule path 'third_party/tensorpipe': checked out '2b4cd91092d335a697416b2a3cb398283246849d' 2025-12-04T08:54:06.6844200Z Submodule path 'third_party/tensorpipe/third_party/googletest': checked out 'aee0f9d9b5b87796ee8a0ab26b7587ec30e8858e' 2025-12-04T08:54:06.6898008Z Submodule path 'third_party/tensorpipe/third_party/libnop': checked out '910b55815be16109f04f4180e9adee14fb4ce281' 2025-12-04T08:54:06.7035138Z Submodule path 'third_party/tensorpipe/third_party/libuv': checked out '5152db2cbfeb5582e9c27c5ea1dba2cd9e10759b' 2025-12-04T08:54:06.7104764Z Submodule path 'third_party/tensorpipe/third_party/pybind11': checked out 'a23996fce38ff6ccfbcdc09f1e63f2c4be5ea2ef' 2025-12-04T08:54:06.7151354Z Submodule path 'third_party/tensorpipe/third_party/pybind11/tools/clang': checked out '6a00cbc4a9b8e68b71caf7f774b3f9c753ae84d5' 2025-12-04T08:54:06.7179262Z [command]/usr/bin/git submodule foreach --recursive git config --local gc.auto 0 2025-12-04T08:54:06.7354158Z Entering 'android/libs/fbjni' 2025-12-04T08:54:06.7373174Z Entering 'third_party/FP16' 2025-12-04T08:54:06.7391942Z Entering 'third_party/FXdiv' 2025-12-04T08:54:06.7411241Z Entering 'third_party/NNPACK' 2025-12-04T08:54:06.7436889Z Entering 'third_party/NVTX' 2025-12-04T08:54:06.7456996Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:06.7475009Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:06.7500563Z Entering 'third_party/aiter' 2025-12-04T08:54:06.7518484Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:06.7541444Z Entering 'third_party/benchmark' 2025-12-04T08:54:06.7561546Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:06.7583293Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:06.7601930Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:06.7624038Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:06.7643380Z Entering 'third_party/cutlass' 2025-12-04T08:54:06.7665030Z Entering 'third_party/fbgemm' 2025-12-04T08:54:06.7685989Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:06.7703664Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:06.7733699Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:06.7752564Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:06.7774688Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:06.7794042Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:06.7814192Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:06.7833817Z Entering 'third_party/flash-attention' 2025-12-04T08:54:06.7855792Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:06.7875824Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:06.7900508Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:06.7925085Z Entering 'third_party/fmt' 2025-12-04T08:54:06.7943787Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:06.7962059Z Entering 'third_party/gloo' 2025-12-04T08:54:06.7980969Z Entering 'third_party/googletest' 2025-12-04T08:54:06.8003349Z Entering 'third_party/ideep' 2025-12-04T08:54:06.8023701Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:06.8045016Z Entering 'third_party/ittapi' 2025-12-04T08:54:06.8064355Z Entering 'third_party/kineto' 2025-12-04T08:54:06.8084355Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:06.8101633Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:06.8121388Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:06.8139948Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:06.8158439Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:06.8178270Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:06.8203340Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:06.8227550Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:06.8250165Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:06.8271057Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:06.8289685Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:06.8308163Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:06.8332601Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:06.8361199Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:06.8387052Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:06.8409333Z Entering 'third_party/kleidiai' 2025-12-04T08:54:06.8429676Z Entering 'third_party/mimalloc' 2025-12-04T08:54:06.8459622Z Entering 'third_party/nlohmann' 2025-12-04T08:54:06.8486104Z Entering 'third_party/onnx' 2025-12-04T08:54:06.8511960Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:06.8537773Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:06.8560120Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:06.8584193Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:06.8604261Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:06.8624450Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:06.8642678Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:06.8660093Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:06.8677816Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:06.8695236Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:06.8714990Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:06.8738493Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:06.8765737Z Entering 'third_party/pocketfft' 2025-12-04T08:54:06.8784617Z Entering 'third_party/protobuf' 2025-12-04T08:54:06.8803884Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:06.8822145Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:06.8842232Z Entering 'third_party/psimd' 2025-12-04T08:54:06.8864401Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:06.8888372Z Entering 'third_party/pybind11' 2025-12-04T08:54:06.8914685Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:06.8934349Z Entering 'third_party/sleef' 2025-12-04T08:54:06.8953933Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:06.8972285Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:06.9002506Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:06.9023032Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:06.9042094Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:06.9062256Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:06.9093995Z ##[endgroup] 2025-12-04T08:54:06.9094192Z ##[group]Persisting credentials for submodules 2025-12-04T08:54:06.9099243Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'url\.https\:\/\/github\.com\/\.insteadOf' && git config --local --unset-all 'url.https://github.com/.insteadOf' || :" 2025-12-04T08:54:06.9249190Z Entering 'android/libs/fbjni' 2025-12-04T08:54:06.9261108Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9261319Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9279807Z Entering 'third_party/FP16' 2025-12-04T08:54:06.9291014Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9291318Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9306651Z Entering 'third_party/FXdiv' 2025-12-04T08:54:06.9318173Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9318434Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9340557Z Entering 'third_party/NNPACK' 2025-12-04T08:54:06.9352530Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9352789Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9370305Z Entering 'third_party/NVTX' 2025-12-04T08:54:06.9383422Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9383641Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9398936Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:06.9415845Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9416033Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9431827Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:06.9444335Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9444519Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9469132Z Entering 'third_party/aiter' 2025-12-04T08:54:06.9481031Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9481203Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9495634Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:06.9507584Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9508474Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9533362Z Entering 'third_party/benchmark' 2025-12-04T08:54:06.9546539Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9546810Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9571676Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:06.9585296Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9585466Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9605013Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:06.9616231Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9616421Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9631950Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:06.9647859Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9648043Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9671771Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:06.9688031Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9688207Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9703924Z Entering 'third_party/cutlass' 2025-12-04T08:54:06.9718762Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9718938Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9738062Z Entering 'third_party/fbgemm' 2025-12-04T08:54:06.9753528Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9753690Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9773849Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:06.9786011Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9786176Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9802454Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:06.9815543Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9815713Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9837842Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:06.9851409Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9851571Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9870950Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:06.9903727Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9903882Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9932692Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:06.9958101Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9958261Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9975891Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:06.9991313Z url.https://github.com/.insteadof 2025-12-04T08:54:06.9991477Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0014918Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:07.0031875Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0032030Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0053944Z Entering 'third_party/flash-attention' 2025-12-04T08:54:07.0071352Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0071652Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0087517Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:07.0099059Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0099187Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0118523Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:07.0129105Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0129239Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0149251Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:07.0161107Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0161243Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0180417Z Entering 'third_party/fmt' 2025-12-04T08:54:07.0195616Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0195765Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0211621Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:07.0226380Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0226527Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0242077Z Entering 'third_party/gloo' 2025-12-04T08:54:07.0258190Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0258333Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0279527Z Entering 'third_party/googletest' 2025-12-04T08:54:07.0291840Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0291973Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0309234Z Entering 'third_party/ideep' 2025-12-04T08:54:07.0321010Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0321136Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0336668Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:07.0356681Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0357030Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0380929Z Entering 'third_party/ittapi' 2025-12-04T08:54:07.0396870Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0396993Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0416059Z Entering 'third_party/kineto' 2025-12-04T08:54:07.0428173Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0428294Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0443922Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:07.0457856Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0457984Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0478128Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:07.0491129Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0491981Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0508246Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:07.0520787Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0520972Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0543882Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:07.0556217Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0556346Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0571857Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:07.0582554Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0582680Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0601251Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:07.0613345Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0613471Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0631990Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:07.0644156Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0644283Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0663428Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:07.0676811Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0677057Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0696534Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:07.0707809Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0708047Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0728803Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:07.0740438Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0740811Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0756805Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:07.0772090Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0772297Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0787898Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:07.0802281Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0802480Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0829917Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:07.0845120Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0845288Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0868089Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:07.0880706Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0880870Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0897340Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:07.0909583Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0909728Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0926720Z Entering 'third_party/kleidiai' 2025-12-04T08:54:07.0941671Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0941825Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0958496Z Entering 'third_party/mimalloc' 2025-12-04T08:54:07.0972720Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0972843Z url.https://github.com/.insteadof 2025-12-04T08:54:07.0991547Z Entering 'third_party/nlohmann' 2025-12-04T08:54:07.1005993Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1006115Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1022967Z Entering 'third_party/onnx' 2025-12-04T08:54:07.1037732Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1037857Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1059858Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:07.1073796Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1073929Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1094569Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:07.1107877Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1108004Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1123359Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:07.1135431Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1135559Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1150870Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:07.1163010Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1163139Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1178097Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:07.1190039Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1190167Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1205788Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:07.1216474Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1216601Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1233527Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:07.1245683Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1245809Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1261731Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:07.1274032Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1274157Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1289514Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:07.1300661Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1300786Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1316371Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:07.1327816Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1327942Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1348963Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:07.1361349Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1361472Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1378951Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:07.1392768Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1392896Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1418618Z Entering 'third_party/pocketfft' 2025-12-04T08:54:07.1432907Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1433032Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1448126Z Entering 'third_party/protobuf' 2025-12-04T08:54:07.1459434Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1459567Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1475626Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:07.1488402Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1488531Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1503988Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:07.1515191Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1515321Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1533255Z Entering 'third_party/psimd' 2025-12-04T08:54:07.1545648Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1545853Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1563243Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:07.1575826Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1575957Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1595007Z Entering 'third_party/pybind11' 2025-12-04T08:54:07.1607737Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1607864Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1625860Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:07.1639296Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1639429Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1655118Z Entering 'third_party/sleef' 2025-12-04T08:54:07.1667286Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1667416Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1682675Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:07.1696202Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1696333Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1711441Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:07.1724907Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1725037Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1742675Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:07.1754939Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1755067Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1773390Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:07.1785915Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1786047Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1801524Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:07.1814040Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1814167Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1828951Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:07.1848842Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1848976Z url.https://github.com/.insteadof 2025-12-04T08:54:07.1879415Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local 'http.https://github.com/.extraheader' 'AUTHORIZATION: basic ***' && git config --local --show-origin --name-only --get-regexp remote.origin.url" 2025-12-04T08:54:07.2039607Z Entering 'android/libs/fbjni' 2025-12-04T08:54:07.2065283Z file:/home/runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T08:54:07.2075275Z Entering 'third_party/FP16' 2025-12-04T08:54:07.2117521Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T08:54:07.2127815Z Entering 'third_party/FXdiv' 2025-12-04T08:54:07.2165634Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T08:54:07.2175741Z Entering 'third_party/NNPACK' 2025-12-04T08:54:07.2194693Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T08:54:07.2204011Z Entering 'third_party/NVTX' 2025-12-04T08:54:07.2224994Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T08:54:07.2236269Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:07.2259043Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T08:54:07.2269919Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:07.2289111Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T08:54:07.2303225Z Entering 'third_party/aiter' 2025-12-04T08:54:07.2323777Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T08:54:07.2333178Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:07.2351536Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T08:54:07.2366411Z Entering 'third_party/benchmark' 2025-12-04T08:54:07.2386923Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:54:07.2399044Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:07.2419292Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T08:54:07.2431142Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:07.2451411Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T08:54:07.2465733Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:07.2504947Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T08:54:07.2515272Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:07.2539578Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T08:54:07.2549044Z Entering 'third_party/cutlass' 2025-12-04T08:54:07.2568795Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T08:54:07.2581619Z Entering 'third_party/fbgemm' 2025-12-04T08:54:07.2600682Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T08:54:07.2609411Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:07.2629120Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T08:54:07.2647604Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:07.2666790Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T08:54:07.2680417Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:07.2697575Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T08:54:07.2706643Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:07.2730731Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T08:54:07.2745898Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:07.2765260Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T08:54:07.2774753Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:07.2793800Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T08:54:07.2802899Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:07.2820690Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T08:54:07.2831951Z Entering 'third_party/flash-attention' 2025-12-04T08:54:07.2850233Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T08:54:07.2860362Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:07.2881062Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T08:54:07.2893020Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:07.2915035Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T08:54:07.2929372Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:07.2948505Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T08:54:07.2959022Z Entering 'third_party/fmt' 2025-12-04T08:54:07.2981778Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T08:54:07.2992954Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:07.3012722Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T08:54:07.3022093Z Entering 'third_party/gloo' 2025-12-04T08:54:07.3044604Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T08:54:07.3054438Z Entering 'third_party/googletest' 2025-12-04T08:54:07.3073424Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:07.3082743Z Entering 'third_party/ideep' 2025-12-04T08:54:07.3099664Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T08:54:07.3108785Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:07.3128368Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T08:54:07.3146735Z Entering 'third_party/ittapi' 2025-12-04T08:54:07.3168210Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T08:54:07.3178263Z Entering 'third_party/kineto' 2025-12-04T08:54:07.3203235Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T08:54:07.3213148Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:07.3233639Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T08:54:07.3243863Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:07.3270532Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T08:54:07.3285884Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:07.3310346Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T08:54:07.3324648Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:07.3344242Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T08:54:07.3352858Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:07.3379291Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T08:54:07.3390007Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:07.3414566Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T08:54:07.3424430Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:07.3450251Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T08:54:07.3459942Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:07.3481015Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:07.3490205Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:07.3519001Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T08:54:07.3528927Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:07.3552214Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T08:54:07.3562452Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:07.3583193Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T08:54:07.3591710Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:07.3617602Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T08:54:07.3628326Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:07.3672880Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T08:54:07.3684971Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:07.3708640Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T08:54:07.3718439Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:07.3739623Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T08:54:07.3754901Z Entering 'third_party/kleidiai' 2025-12-04T08:54:07.3775240Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T08:54:07.3787974Z Entering 'third_party/mimalloc' 2025-12-04T08:54:07.3809243Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T08:54:07.3819568Z Entering 'third_party/nlohmann' 2025-12-04T08:54:07.3841669Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T08:54:07.3851222Z Entering 'third_party/onnx' 2025-12-04T08:54:07.3872063Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T08:54:07.3894357Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:07.3915399Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:54:07.3927343Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:07.3946981Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T08:54:07.3961960Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:07.3983279Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:54:07.3992917Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:07.4013881Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:07.4023385Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:07.4045269Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T08:54:07.4055127Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:07.4074225Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T08:54:07.4087145Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:07.4106622Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T08:54:07.4115721Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:07.4135317Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T08:54:07.4143368Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:07.4162933Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T08:54:07.4171665Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:07.4211223Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T08:54:07.4222295Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:07.4246676Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T08:54:07.4257641Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:07.4276432Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T08:54:07.4293455Z Entering 'third_party/pocketfft' 2025-12-04T08:54:07.4312808Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T08:54:07.4321888Z Entering 'third_party/protobuf' 2025-12-04T08:54:07.4340767Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T08:54:07.4352439Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:07.4376477Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T08:54:07.4390280Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:07.4409843Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:07.4421282Z Entering 'third_party/psimd' 2025-12-04T08:54:07.4442421Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T08:54:07.4452186Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:07.4472946Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T08:54:07.4482596Z Entering 'third_party/pybind11' 2025-12-04T08:54:07.4501762Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:54:07.4511425Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:07.4530249Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T08:54:07.4543469Z Entering 'third_party/sleef' 2025-12-04T08:54:07.4562187Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T08:54:07.4571728Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:07.4589594Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T08:54:07.4599136Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:07.4623777Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T08:54:07.4634071Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:07.4652962Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T08:54:07.4667570Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:07.4686834Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T08:54:07.4696059Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:07.4713897Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T08:54:07.4722564Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:07.4743660Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T08:54:07.4948197Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'git@github.com:' 2025-12-04T08:54:07.5097437Z Entering 'android/libs/fbjni' 2025-12-04T08:54:07.5130551Z Entering 'third_party/FP16' 2025-12-04T08:54:07.5149827Z Entering 'third_party/FXdiv' 2025-12-04T08:54:07.5170554Z Entering 'third_party/NNPACK' 2025-12-04T08:54:07.5195733Z Entering 'third_party/NVTX' 2025-12-04T08:54:07.5218094Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:07.5237509Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:07.5261682Z Entering 'third_party/aiter' 2025-12-04T08:54:07.5280884Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:07.5302162Z Entering 'third_party/benchmark' 2025-12-04T08:54:07.5319472Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:07.5342071Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:07.5360773Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:07.5378771Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:07.5396580Z Entering 'third_party/cutlass' 2025-12-04T08:54:07.5419852Z Entering 'third_party/fbgemm' 2025-12-04T08:54:07.5439775Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:07.5457116Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:07.5478259Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:07.5496191Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:07.5515711Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:07.5533052Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:07.5551016Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:07.5571473Z Entering 'third_party/flash-attention' 2025-12-04T08:54:07.5598698Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:07.5618776Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:07.5649888Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:07.5669585Z Entering 'third_party/fmt' 2025-12-04T08:54:07.5690238Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:07.5710836Z Entering 'third_party/gloo' 2025-12-04T08:54:07.5730921Z Entering 'third_party/googletest' 2025-12-04T08:54:07.5756019Z Entering 'third_party/ideep' 2025-12-04T08:54:07.5776008Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:07.5808292Z Entering 'third_party/ittapi' 2025-12-04T08:54:07.5827626Z Entering 'third_party/kineto' 2025-12-04T08:54:07.5846131Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:07.5863199Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:07.5882683Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:07.5902091Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:07.5923465Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:07.5943466Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:07.5963310Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:07.5983843Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:07.6002240Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:07.6021299Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:07.6038343Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:07.6062454Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:07.6091218Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:07.6122918Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:07.6149363Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:07.6171519Z Entering 'third_party/kleidiai' 2025-12-04T08:54:07.6191115Z Entering 'third_party/mimalloc' 2025-12-04T08:54:07.6211594Z Entering 'third_party/nlohmann' 2025-12-04T08:54:07.6229989Z Entering 'third_party/onnx' 2025-12-04T08:54:07.6253745Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:07.6276858Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:07.6295336Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:07.6313064Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:07.6331095Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:07.6350804Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:07.6369516Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:07.6387636Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:07.6404109Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:07.6426475Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:07.6444459Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:07.6472186Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:07.6497648Z Entering 'third_party/pocketfft' 2025-12-04T08:54:07.6516347Z Entering 'third_party/protobuf' 2025-12-04T08:54:07.6537167Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:07.6569699Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:07.6599123Z Entering 'third_party/psimd' 2025-12-04T08:54:07.6623272Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:07.6642218Z Entering 'third_party/pybind11' 2025-12-04T08:54:07.6661212Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:07.6681748Z Entering 'third_party/sleef' 2025-12-04T08:54:07.6700546Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:07.6721397Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:07.6740088Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:07.6759407Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:07.6779366Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:07.6798454Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:07.6828929Z [command]/usr/bin/git submodule foreach --recursive git config --local --add 'url.https://github.com/.insteadOf' 'org-21003710@github.com:' 2025-12-04T08:54:07.6983125Z Entering 'android/libs/fbjni' 2025-12-04T08:54:07.7001336Z Entering 'third_party/FP16' 2025-12-04T08:54:07.7026568Z Entering 'third_party/FXdiv' 2025-12-04T08:54:07.7045297Z Entering 'third_party/NNPACK' 2025-12-04T08:54:07.7064301Z Entering 'third_party/NVTX' 2025-12-04T08:54:07.7084146Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T08:54:07.7103332Z Entering 'third_party/XNNPACK' 2025-12-04T08:54:07.7133076Z Entering 'third_party/aiter' 2025-12-04T08:54:07.7152221Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T08:54:07.7174449Z Entering 'third_party/benchmark' 2025-12-04T08:54:07.7197723Z Entering 'third_party/composable_kernel' 2025-12-04T08:54:07.7219721Z Entering 'third_party/cpp-httplib' 2025-12-04T08:54:07.7238438Z Entering 'third_party/cpuinfo' 2025-12-04T08:54:07.7257259Z Entering 'third_party/cudnn_frontend' 2025-12-04T08:54:07.7275190Z Entering 'third_party/cutlass' 2025-12-04T08:54:07.7295877Z Entering 'third_party/fbgemm' 2025-12-04T08:54:07.7315984Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T08:54:07.7332975Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T08:54:07.7354360Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T08:54:07.7374123Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T08:54:07.7394301Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T08:54:07.7414783Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T08:54:07.7440085Z Entering 'third_party/fbgemm/external/json' 2025-12-04T08:54:07.7460423Z Entering 'third_party/flash-attention' 2025-12-04T08:54:07.7480231Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T08:54:07.7499562Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T08:54:07.7520794Z Entering 'third_party/flatbuffers' 2025-12-04T08:54:07.7541274Z Entering 'third_party/fmt' 2025-12-04T08:54:07.7565073Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T08:54:07.7587783Z Entering 'third_party/gloo' 2025-12-04T08:54:07.7607051Z Entering 'third_party/googletest' 2025-12-04T08:54:07.7630711Z Entering 'third_party/ideep' 2025-12-04T08:54:07.7649013Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T08:54:07.7669492Z Entering 'third_party/ittapi' 2025-12-04T08:54:07.7689311Z Entering 'third_party/kineto' 2025-12-04T08:54:07.7709190Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T08:54:07.7736292Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T08:54:07.7759237Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T08:54:07.7780077Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T08:54:07.7800370Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T08:54:07.7821512Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T08:54:07.7840760Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T08:54:07.7858742Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T08:54:07.7880313Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T08:54:07.7898932Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T08:54:07.7917491Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T08:54:07.7935071Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:07.7954108Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:07.7976881Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T08:54:07.7994920Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T08:54:07.8013531Z Entering 'third_party/kleidiai' 2025-12-04T08:54:07.8032778Z Entering 'third_party/mimalloc' 2025-12-04T08:54:07.8056268Z Entering 'third_party/nlohmann' 2025-12-04T08:54:07.8077390Z Entering 'third_party/onnx' 2025-12-04T08:54:07.8103272Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T08:54:07.8129443Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T08:54:07.8157203Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T08:54:07.8182723Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T08:54:07.8202754Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T08:54:07.8222258Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T08:54:07.8246500Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T08:54:07.8265454Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T08:54:07.8287216Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T08:54:07.8304433Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T08:54:07.8327763Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T08:54:07.8349291Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T08:54:07.8382296Z Entering 'third_party/pocketfft' 2025-12-04T08:54:07.8407416Z Entering 'third_party/protobuf' 2025-12-04T08:54:07.8431319Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T08:54:07.8451194Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T08:54:07.8479649Z Entering 'third_party/psimd' 2025-12-04T08:54:07.8498896Z Entering 'third_party/pthreadpool' 2025-12-04T08:54:07.8518884Z Entering 'third_party/pybind11' 2025-12-04T08:54:07.8539096Z Entering 'third_party/python-peachpy' 2025-12-04T08:54:07.8558041Z Entering 'third_party/sleef' 2025-12-04T08:54:07.8579536Z Entering 'third_party/tensorpipe' 2025-12-04T08:54:07.8601153Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T08:54:07.8623265Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T08:54:07.8642099Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T08:54:07.8661219Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T08:54:07.8680726Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T08:54:07.8714947Z ##[endgroup] 2025-12-04T08:54:07.8953906Z [command]/usr/bin/git log -1 --format=%H 2025-12-04T08:54:07.9071736Z ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:54:07.9203023Z Prepare all required actions 2025-12-04T08:54:07.9203296Z Getting action download info 2025-12-04T08:54:08.1833249Z Download action repository 'aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076' (SHA:062b18b96a7aff071d4dc91bc00c4c1a7945b076) 2025-12-04T08:54:08.9145775Z ##[group]Run ./.github/actions/setup-rocm 2025-12-04T08:54:08.9145916Z env: 2025-12-04T08:54:08.9146006Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:08.9146109Z ##[endgroup] 2025-12-04T08:54:08.9160112Z ##[group]Run dpkg -l | grep -E " rocm" 2025-12-04T08:54:08.9160292Z dpkg -l | grep -E " rocm" 2025-12-04T08:54:08.9163934Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:08.9164083Z env: 2025-12-04T08:54:08.9164171Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:08.9164275Z ##[endgroup] 2025-12-04T08:54:08.9230234Z ii rocm-cmake 0.14.0.60401-83~22.04 amd64 rocm-cmake built using CMake 2025-12-04T08:54:08.9231149Z ii rocm-core 6.4.1.60401-83~22.04 amd64 ROCm Runtime software stack 2025-12-04T08:54:08.9231631Z ii rocm-dbgapi 0.77.2.60401-83~22.04 amd64 Library to provide AMD GPU debugger API 2025-12-04T08:54:08.9232202Z ii rocm-debug-agent 2.0.4.60401-83~22.04 amd64 Radeon Open Compute Debug Agent (ROCdebug-agent) 2025-12-04T08:54:08.9232788Z ii rocm-dev 6.4.1.60401-83~22.04 amd64 Radeon Open Compute (ROCm) Runtime software stack 2025-12-04T08:54:08.9233306Z ii rocm-device-libs 1.0.0.60401-83~22.04 amd64 Radeon Open Compute - device libraries 2025-12-04T08:54:08.9233785Z ii rocm-gdb 15.2.60401-83~22.04 amd64 ROCgdb 2025-12-04T08:54:08.9234194Z ii rocm-llvm 19.0.0.25184.60401-83~22.04 amd64 ROCm core compiler 2025-12-04T08:54:08.9234648Z ii rocm-opencl 2.0.0.60401-83~22.04 amd64 clr built using CMake 2025-12-04T08:54:08.9235093Z ii rocm-opencl-dev 2.0.0.60401-83~22.04 amd64 clr built using CMake 2025-12-04T08:54:08.9235544Z ii rocm-smi-lib 7.5.0.60401-83~22.04 amd64 AMD System Management libraries 2025-12-04T08:54:08.9236025Z ii rocm-utils 6.4.1.60401-83~22.04 amd64 Radeon Open Compute (ROCm) Runtime software stack 2025-12-04T08:54:08.9236530Z ii rocminfo 1.0.0.60401-83~22.04 amd64 Radeon Open Compute (ROCm) Runtime rocminfo tool 2025-12-04T08:54:08.9254749Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T08:54:08.9255044Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T08:54:08.9255229Z # shellcheck disable=SC2046 2025-12-04T08:54:08.9255383Z docker stop $(docker ps -q) || true 2025-12-04T08:54:08.9255531Z # Prune all stopped containers. 2025-12-04T08:54:08.9255673Z docker container prune -f 2025-12-04T08:54:08.9260749Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:08.9260915Z env: 2025-12-04T08:54:08.9261024Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:08.9261148Z ##[endgroup] 2025-12-04T08:54:08.9476939Z docker: 'docker stop' requires at least 1 argument 2025-12-04T08:54:08.9477190Z 2025-12-04T08:54:08.9477333Z Usage: docker stop [OPTIONS] CONTAINER [CONTAINER...] 2025-12-04T08:54:08.9477541Z 2025-12-04T08:54:08.9477658Z See 'docker stop --help' for more information 2025-12-04T08:54:08.9581214Z Total reclaimed space: 0B 2025-12-04T08:54:08.9608573Z ##[group]Run cat /etc/os-release || true 2025-12-04T08:54:08.9608775Z cat /etc/os-release || true 2025-12-04T08:54:08.9608976Z cat /etc/apt/sources.list.d/rocm.list || true 2025-12-04T08:54:08.9609333Z cat /opt/rocm/.info/version || true 2025-12-04T08:54:08.9609485Z whoami 2025-12-04T08:54:08.9614628Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:08.9614831Z env: 2025-12-04T08:54:08.9614994Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:08.9615103Z ##[endgroup] 2025-12-04T08:54:08.9642797Z PRETTY_NAME="Ubuntu 22.04.5 LTS" 2025-12-04T08:54:08.9643004Z NAME="Ubuntu" 2025-12-04T08:54:08.9643102Z VERSION_ID="22.04" 2025-12-04T08:54:08.9643216Z VERSION="22.04.5 LTS (Jammy Jellyfish)" 2025-12-04T08:54:08.9643343Z VERSION_CODENAME=jammy 2025-12-04T08:54:08.9643449Z ID=ubuntu 2025-12-04T08:54:08.9643538Z ID_LIKE=debian 2025-12-04T08:54:08.9643669Z HOME_URL="https://www.ubuntu.com/" 2025-12-04T08:54:08.9643807Z SUPPORT_URL="https://help.ubuntu.com/" 2025-12-04T08:54:08.9643974Z BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2025-12-04T08:54:08.9644199Z PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2025-12-04T08:54:08.9645977Z UBUNTU_CODENAME=jammy 2025-12-04T08:54:08.9649371Z deb [arch=amd64 signed-by=/etc/apt/keyrings/rocm.gpg] https://repo.radeon.com/rocm/apt/6.4.1 jammy main 2025-12-04T08:54:08.9655914Z 6.4.1-83 2025-12-04T08:54:08.9661363Z runner 2025-12-04T08:54:08.9679695Z ##[group]Run dpkg -l | grep -E " amdgpu" 2025-12-04T08:54:08.9679890Z dpkg -l | grep -E " amdgpu" 2025-12-04T08:54:08.9684033Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:08.9684187Z env: 2025-12-04T08:54:08.9684285Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:08.9684394Z ##[endgroup] 2025-12-04T08:54:08.9736872Z ii amdgpu-core 1:6.4.60401-2164967.22.04 all Core meta package for unified amdgpu driver. 2025-12-04T08:54:08.9737128Z ii amdgpu-install 6.4.60401-2164967.22.04 all AMDGPU driver repository and installer 2025-12-04T08:54:08.9759710Z ##[group]Run rocm-smi 2025-12-04T08:54:08.9759935Z rocm-smi 2025-12-04T08:54:08.9764887Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:08.9765078Z env: 2025-12-04T08:54:08.9765186Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:08.9765315Z ##[endgroup] 2025-12-04T08:54:09.0317186Z 2025-12-04T08:54:09.0317333Z 2025-12-04T08:54:09.0317485Z ============================================ ROCm System Management Interface ============================================ 2025-12-04T08:54:09.0317703Z ====================================================== Concise Info ====================================================== 2025-12-04T08:54:09.0317934Z Device Node IDs Temp Power Partitions SCLK MCLK Fan Perf PwrCap VRAM% GPU% 2025-12-04T08:54:09.0318485Z  (DID, GUID) (Junction) (Socket) (Mem, Compute, ID)  2025-12-04T08:54:09.0318687Z ========================================================================================================================== 2025-12-04T08:54:09.0319113Z 0 5 0x74a5, 2987 28.0°C 114.0W NPS1, SPX, 0 N/A 900Mhz 0% manual 1000.0W 0% 0% 2025-12-04T08:54:09.0319869Z ========================================================================================================================== 2025-12-04T08:54:09.0320046Z ================================================== End of ROCm SMI Log =================================================== 2025-12-04T08:54:09.0381089Z ##[group]Run rocminfo 2025-12-04T08:54:09.0381232Z rocminfo 2025-12-04T08:54:09.0384882Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:09.0385026Z env: 2025-12-04T08:54:09.0385112Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:09.0385215Z ##[endgroup] 2025-12-04T08:54:09.0904501Z ROCk module version 6.12.12 is loaded 2025-12-04T08:54:09.0904633Z ===================== 2025-12-04T08:54:09.0904758Z HSA System Attributes 2025-12-04T08:54:09.0904862Z ===================== 2025-12-04T08:54:09.0905208Z Runtime Version: 1.15 2025-12-04T08:54:09.0905321Z Runtime Ext Version: 1.7 2025-12-04T08:54:09.0905436Z System Timestamp Freq.: 1000.000000MHz 2025-12-04T08:54:09.0905614Z Sig. Max Wait Duration: 18446744073709551615 (0xFFFFFFFFFFFFFFFF) (timestamp count) 2025-12-04T08:54:09.0905800Z Machine Model: LARGE 2025-12-04T08:54:09.0905956Z System Endianness: LITTLE 2025-12-04T08:54:09.0906096Z Mwaitx: DISABLED 2025-12-04T08:54:09.0906209Z XNACK enabled: NO 2025-12-04T08:54:09.0906317Z DMAbuf Support: YES 2025-12-04T08:54:09.0906424Z VMM Support: YES 2025-12-04T08:54:09.0906492Z 2025-12-04T08:54:09.0906529Z ========== 2025-12-04T08:54:09.0906623Z HSA Agents 2025-12-04T08:54:09.0906716Z ========== 2025-12-04T08:54:09.0907008Z ******* 2025-12-04T08:54:09.0907098Z Agent 1 2025-12-04T08:54:09.0907196Z ******* 2025-12-04T08:54:09.0907310Z Name: AMD EPYC 9575F 64-Core Processor 2025-12-04T08:54:09.0907492Z Uuid: CPU-XX 2025-12-04T08:54:09.0907642Z Marketing Name: AMD EPYC 9575F 64-Core Processor 2025-12-04T08:54:09.0907800Z Vendor Name: CPU 2025-12-04T08:54:09.0907946Z Feature: None specified 2025-12-04T08:54:09.0908096Z Profile: FULL_PROFILE 2025-12-04T08:54:09.0908241Z Float Round Mode: NEAR 2025-12-04T08:54:09.0908392Z Max Queue Number: 0(0x0) 2025-12-04T08:54:09.0908540Z Queue Min Size: 0(0x0) 2025-12-04T08:54:09.0908686Z Queue Max Size: 0(0x0) 2025-12-04T08:54:09.0908850Z Queue Type: MULTI 2025-12-04T08:54:09.0908990Z Node: 0 2025-12-04T08:54:09.0909129Z Device Type: CPU 2025-12-04T08:54:09.0909259Z Cache Info: 2025-12-04T08:54:09.0909371Z L1: 49152(0xc000) KB 2025-12-04T08:54:09.0909505Z Chip ID: 0(0x0) 2025-12-04T08:54:09.0909645Z ASIC Revision: 0(0x0) 2025-12-04T08:54:09.0909794Z Cacheline Size: 64(0x40) 2025-12-04T08:54:09.0909947Z Max Clock Freq. (MHz): 3300 2025-12-04T08:54:09.0910094Z BDFID: 0 2025-12-04T08:54:09.0910234Z Internal Node ID: 0 2025-12-04T08:54:09.0910385Z Compute Unit: 64 2025-12-04T08:54:09.0910531Z SIMDs per CU: 0 2025-12-04T08:54:09.0910677Z Shader Engines: 0 2025-12-04T08:54:09.0910828Z Shader Arrs. per Eng.: 0 2025-12-04T08:54:09.0910981Z WatchPts on Addr. Ranges:1 2025-12-04T08:54:09.0911167Z Memory Properties: 2025-12-04T08:54:09.0911272Z Features: None 2025-12-04T08:54:09.0911376Z Pool Info: 2025-12-04T08:54:09.0911474Z Pool 1 2025-12-04T08:54:09.0911603Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-12-04T08:54:09.0911753Z Size: 1584734448(0x5e7520f0) KB 2025-12-04T08:54:09.0911896Z Allocatable: TRUE 2025-12-04T08:54:09.0923987Z Alloc Granule: 4KB 2025-12-04T08:54:09.0924267Z Alloc Recommended Granule:4KB 2025-12-04T08:54:09.0924433Z Alloc Alignment: 4KB 2025-12-04T08:54:09.0924596Z Accessible by all: TRUE 2025-12-04T08:54:09.0924734Z Pool 2 2025-12-04T08:54:09.0924864Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-12-04T08:54:09.0925008Z Size: 1584734448(0x5e7520f0) KB 2025-12-04T08:54:09.0925157Z Allocatable: TRUE 2025-12-04T08:54:09.0925316Z Alloc Granule: 4KB 2025-12-04T08:54:09.0925476Z Alloc Recommended Granule:4KB 2025-12-04T08:54:09.0925640Z Alloc Alignment: 4KB 2025-12-04T08:54:09.0925798Z Accessible by all: TRUE 2025-12-04T08:54:09.0925988Z Pool 3 2025-12-04T08:54:09.0926120Z Segment: GLOBAL; FLAGS: KERNARG, FINE GRAINED 2025-12-04T08:54:09.0926266Z Size: 1584734448(0x5e7520f0) KB 2025-12-04T08:54:09.0926414Z Allocatable: TRUE 2025-12-04T08:54:09.0926571Z Alloc Granule: 4KB 2025-12-04T08:54:09.0926729Z Alloc Recommended Granule:4KB 2025-12-04T08:54:09.0926951Z Alloc Alignment: 4KB 2025-12-04T08:54:09.0927110Z Accessible by all: TRUE 2025-12-04T08:54:09.0927245Z Pool 4 2025-12-04T08:54:09.0927375Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-12-04T08:54:09.0927526Z Size: 1584734448(0x5e7520f0) KB 2025-12-04T08:54:09.0927681Z Allocatable: TRUE 2025-12-04T08:54:09.0927846Z Alloc Granule: 4KB 2025-12-04T08:54:09.0928008Z Alloc Recommended Granule:4KB 2025-12-04T08:54:09.0928178Z Alloc Alignment: 4KB 2025-12-04T08:54:09.0928338Z Accessible by all: TRUE 2025-12-04T08:54:09.0928478Z ISA Info: 2025-12-04T08:54:09.0928585Z ******* 2025-12-04T08:54:09.0928694Z Agent 2 2025-12-04T08:54:09.0928797Z ******* 2025-12-04T08:54:09.0928926Z Name: AMD EPYC 9575F 64-Core Processor 2025-12-04T08:54:09.0929073Z Uuid: CPU-XX 2025-12-04T08:54:09.0929237Z Marketing Name: AMD EPYC 9575F 64-Core Processor 2025-12-04T08:54:09.0929405Z Vendor Name: CPU 2025-12-04T08:54:09.0929565Z Feature: None specified 2025-12-04T08:54:09.0929727Z Profile: FULL_PROFILE 2025-12-04T08:54:09.0929889Z Float Round Mode: NEAR 2025-12-04T08:54:09.0930044Z Max Queue Number: 0(0x0) 2025-12-04T08:54:09.0930209Z Queue Min Size: 0(0x0) 2025-12-04T08:54:09.0930360Z Queue Max Size: 0(0x0) 2025-12-04T08:54:09.0930518Z Queue Type: MULTI 2025-12-04T08:54:09.0930675Z Node: 1 2025-12-04T08:54:09.0930820Z Device Type: CPU 2025-12-04T08:54:09.0930963Z Cache Info: 2025-12-04T08:54:09.0931087Z L1: 49152(0xc000) KB 2025-12-04T08:54:09.0931279Z Chip ID: 0(0x0) 2025-12-04T08:54:09.0931434Z ASIC Revision: 0(0x0) 2025-12-04T08:54:09.0931590Z Cacheline Size: 64(0x40) 2025-12-04T08:54:09.0931761Z Max Clock Freq. (MHz): 3300 2025-12-04T08:54:09.0931918Z BDFID: 0 2025-12-04T08:54:09.0932067Z Internal Node ID: 1 2025-12-04T08:54:09.0932231Z Compute Unit: 64 2025-12-04T08:54:09.0932392Z SIMDs per CU: 0 2025-12-04T08:54:09.0932547Z Shader Engines: 0 2025-12-04T08:54:09.0932711Z Shader Arrs. per Eng.: 0 2025-12-04T08:54:09.0932871Z WatchPts on Addr. Ranges:1 2025-12-04T08:54:09.0933057Z Memory Properties: 2025-12-04T08:54:09.0933186Z Features: None 2025-12-04T08:54:09.0933296Z Pool Info: 2025-12-04T08:54:09.0933409Z Pool 1 2025-12-04T08:54:09.0933544Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-12-04T08:54:09.0933694Z Size: 1585355632(0x5e7e9b70) KB 2025-12-04T08:54:09.0933853Z Allocatable: TRUE 2025-12-04T08:54:09.0934010Z Alloc Granule: 4KB 2025-12-04T08:54:09.0934181Z Alloc Recommended Granule:4KB 2025-12-04T08:54:09.0934348Z Alloc Alignment: 4KB 2025-12-04T08:54:09.0934505Z Accessible by all: TRUE 2025-12-04T08:54:09.0934638Z Pool 2 2025-12-04T08:54:09.0934764Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-12-04T08:54:09.0934913Z Size: 1585355632(0x5e7e9b70) KB 2025-12-04T08:54:09.0935065Z Allocatable: TRUE 2025-12-04T08:54:09.0935221Z Alloc Granule: 4KB 2025-12-04T08:54:09.0935390Z Alloc Recommended Granule:4KB 2025-12-04T08:54:09.0935559Z Alloc Alignment: 4KB 2025-12-04T08:54:09.0935713Z Accessible by all: TRUE 2025-12-04T08:54:09.0935855Z Pool 3 2025-12-04T08:54:09.0935987Z Segment: GLOBAL; FLAGS: KERNARG, FINE GRAINED 2025-12-04T08:54:09.0936133Z Size: 1585355632(0x5e7e9b70) KB 2025-12-04T08:54:09.0936280Z Allocatable: TRUE 2025-12-04T08:54:09.0936433Z Alloc Granule: 4KB 2025-12-04T08:54:09.0936594Z Alloc Recommended Granule:4KB 2025-12-04T08:54:09.0936809Z Alloc Alignment: 4KB 2025-12-04T08:54:09.0936960Z Accessible by all: TRUE 2025-12-04T08:54:09.0937094Z Pool 4 2025-12-04T08:54:09.0937214Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-12-04T08:54:09.0937351Z Size: 1585355632(0x5e7e9b70) KB 2025-12-04T08:54:09.0937491Z Allocatable: TRUE 2025-12-04T08:54:09.0937647Z Alloc Granule: 4KB 2025-12-04T08:54:09.0937799Z Alloc Recommended Granule:4KB 2025-12-04T08:54:09.0937953Z Alloc Alignment: 4KB 2025-12-04T08:54:09.0938103Z Accessible by all: TRUE 2025-12-04T08:54:09.0938237Z ISA Info: 2025-12-04T08:54:09.0938392Z ******* 2025-12-04T08:54:09.0938486Z Agent 3 2025-12-04T08:54:09.0938578Z ******* 2025-12-04T08:54:09.0938685Z Name: gfx942 2025-12-04T08:54:09.0938821Z Uuid: GPU-8052c781c18bf05c 2025-12-04T08:54:09.0938968Z Marketing Name: AMD Instinct MI325X 2025-12-04T08:54:09.0939115Z Vendor Name: AMD 2025-12-04T08:54:09.0939262Z Feature: KERNEL_DISPATCH 2025-12-04T08:54:09.0939405Z Profile: BASE_PROFILE 2025-12-04T08:54:09.0939550Z Float Round Mode: NEAR 2025-12-04T08:54:09.0939697Z Max Queue Number: 128(0x80) 2025-12-04T08:54:09.0939842Z Queue Min Size: 64(0x40) 2025-12-04T08:54:09.0940030Z Queue Max Size: 131072(0x20000) 2025-12-04T08:54:09.0940173Z Queue Type: MULTI 2025-12-04T08:54:09.0940307Z Node: 2 2025-12-04T08:54:09.0940445Z Device Type: GPU 2025-12-04T08:54:09.0940571Z Cache Info: 2025-12-04T08:54:09.0940679Z L1: 32(0x20) KB 2025-12-04T08:54:09.0940806Z L2: 4096(0x1000) KB 2025-12-04T08:54:09.0940933Z L3: 262144(0x40000) KB 2025-12-04T08:54:09.0941060Z Chip ID: 29861(0x74a5) 2025-12-04T08:54:09.0941203Z ASIC Revision: 1(0x1) 2025-12-04T08:54:09.0941347Z Cacheline Size: 128(0x80) 2025-12-04T08:54:09.0941502Z Max Clock Freq. (MHz): 2100 2025-12-04T08:54:09.0941643Z BDFID: 5376 2025-12-04T08:54:09.0941782Z Internal Node ID: 2 2025-12-04T08:54:09.0941926Z Compute Unit: 304 2025-12-04T08:54:09.0942068Z SIMDs per CU: 4 2025-12-04T08:54:09.0942208Z Shader Engines: 32 2025-12-04T08:54:09.0942357Z Shader Arrs. per Eng.: 1 2025-12-04T08:54:09.0942510Z WatchPts on Addr. Ranges:4 2025-12-04T08:54:09.0942664Z Coherent Host Access: FALSE 2025-12-04T08:54:09.0942797Z Memory Properties: 2025-12-04T08:54:09.0942906Z Features: KERNEL_DISPATCH 2025-12-04T08:54:09.0943043Z Fast F16 Operation: TRUE 2025-12-04T08:54:09.0943199Z Wavefront Size: 64(0x40) 2025-12-04T08:54:09.0943346Z Workgroup Max Size: 1024(0x400) 2025-12-04T08:54:09.0943484Z Workgroup Max Size per Dimension: 2025-12-04T08:54:09.0943605Z x 1024(0x400) 2025-12-04T08:54:09.0943727Z y 1024(0x400) 2025-12-04T08:54:09.0943847Z z 1024(0x400) 2025-12-04T08:54:09.0943979Z Max Waves Per CU: 32(0x20) 2025-12-04T08:54:09.0944127Z Max Work-item Per CU: 2048(0x800) 2025-12-04T08:54:09.0944276Z Grid Max Size: 4294967295(0xffffffff) 2025-12-04T08:54:09.0944405Z Grid Max Size per Dimension: 2025-12-04T08:54:09.0944514Z x 4294967295(0xffffffff) 2025-12-04T08:54:09.0944640Z y 4294967295(0xffffffff) 2025-12-04T08:54:09.0944793Z z 4294967295(0xffffffff) 2025-12-04T08:54:09.0944937Z Max fbarriers/Workgrp: 32 2025-12-04T08:54:09.0949686Z Packet Processor uCode:: 185 2025-12-04T08:54:09.0949851Z SDMA engine uCode:: 24 2025-12-04T08:54:09.0950004Z IOMMU Support:: None 2025-12-04T08:54:09.0950135Z Pool Info: 2025-12-04T08:54:09.0950235Z Pool 1 2025-12-04T08:54:09.0950366Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-12-04T08:54:09.0950514Z Size: 268419072(0xfffc000) KB 2025-12-04T08:54:09.0950656Z Allocatable: TRUE 2025-12-04T08:54:09.0951015Z Alloc Granule: 4KB 2025-12-04T08:54:09.0951338Z Alloc Recommended Granule:2048KB 2025-12-04T08:54:09.0951492Z Alloc Alignment: 4KB 2025-12-04T08:54:09.0951664Z Accessible by all: FALSE 2025-12-04T08:54:09.0951797Z Pool 2 2025-12-04T08:54:09.0951920Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-12-04T08:54:09.0952069Z Size: 268419072(0xfffc000) KB 2025-12-04T08:54:09.0952210Z Allocatable: TRUE 2025-12-04T08:54:09.0952360Z Alloc Granule: 4KB 2025-12-04T08:54:09.0952514Z Alloc Recommended Granule:2048KB 2025-12-04T08:54:09.0952665Z Alloc Alignment: 4KB 2025-12-04T08:54:09.0952817Z Accessible by all: FALSE 2025-12-04T08:54:09.0952952Z Pool 3 2025-12-04T08:54:09.0953073Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-12-04T08:54:09.0953211Z Size: 268419072(0xfffc000) KB 2025-12-04T08:54:09.0953349Z Allocatable: TRUE 2025-12-04T08:54:09.0953499Z Alloc Granule: 4KB 2025-12-04T08:54:09.0953657Z Alloc Recommended Granule:2048KB 2025-12-04T08:54:09.0953808Z Alloc Alignment: 4KB 2025-12-04T08:54:09.0953958Z Accessible by all: FALSE 2025-12-04T08:54:09.0954088Z Pool 4 2025-12-04T08:54:09.0954202Z Segment: GROUP 2025-12-04T08:54:09.0954338Z Size: 64(0x40) KB 2025-12-04T08:54:09.0954478Z Allocatable: FALSE 2025-12-04T08:54:09.0954629Z Alloc Granule: 0KB 2025-12-04T08:54:09.0954783Z Alloc Recommended Granule:0KB 2025-12-04T08:54:09.0954935Z Alloc Alignment: 0KB 2025-12-04T08:54:09.0955085Z Accessible by all: FALSE 2025-12-04T08:54:09.0955215Z ISA Info: 2025-12-04T08:54:09.0955311Z ISA 1 2025-12-04T08:54:09.0955437Z Name: amdgcn-amd-amdhsa--gfx942:sramecc+:xnack- 2025-12-04T08:54:09.0955595Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-12-04T08:54:09.0955748Z Profiles: HSA_PROFILE_BASE 2025-12-04T08:54:09.0955902Z Default Rounding Mode: NEAR 2025-12-04T08:54:09.0956056Z Default Rounding Mode: NEAR 2025-12-04T08:54:09.0956255Z Fast f16: TRUE 2025-12-04T08:54:09.0956401Z Workgroup Max Size: 1024(0x400) 2025-12-04T08:54:09.0956538Z Workgroup Max Size per Dimension: 2025-12-04T08:54:09.0956664Z x 1024(0x400) 2025-12-04T08:54:09.0956840Z y 1024(0x400) 2025-12-04T08:54:09.0956961Z z 1024(0x400) 2025-12-04T08:54:09.0957107Z Grid Max Size: 4294967295(0xffffffff) 2025-12-04T08:54:09.0957241Z Grid Max Size per Dimension: 2025-12-04T08:54:09.0957361Z x 4294967295(0xffffffff) 2025-12-04T08:54:09.0957486Z y 4294967295(0xffffffff) 2025-12-04T08:54:09.0957614Z z 4294967295(0xffffffff) 2025-12-04T08:54:09.0957800Z FBarrier Max Size: 32 2025-12-04T08:54:09.0957930Z ISA 2 2025-12-04T08:54:09.0958069Z Name: amdgcn-amd-amdhsa--gfx9-4-generic:sramecc+:xnack- 2025-12-04T08:54:09.0958243Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-12-04T08:54:09.0958398Z Profiles: HSA_PROFILE_BASE 2025-12-04T08:54:09.0958556Z Default Rounding Mode: NEAR 2025-12-04T08:54:09.0958713Z Default Rounding Mode: NEAR 2025-12-04T08:54:09.0958863Z Fast f16: TRUE 2025-12-04T08:54:09.0959014Z Workgroup Max Size: 1024(0x400) 2025-12-04T08:54:09.0959152Z Workgroup Max Size per Dimension: 2025-12-04T08:54:09.0959279Z x 1024(0x400) 2025-12-04T08:54:09.0959410Z y 1024(0x400) 2025-12-04T08:54:09.0959535Z z 1024(0x400) 2025-12-04T08:54:09.0959672Z Grid Max Size: 4294967295(0xffffffff) 2025-12-04T08:54:09.0959808Z Grid Max Size per Dimension: 2025-12-04T08:54:09.0959923Z x 4294967295(0xffffffff) 2025-12-04T08:54:09.0960053Z y 4294967295(0xffffffff) 2025-12-04T08:54:09.0960178Z z 4294967295(0xffffffff) 2025-12-04T08:54:09.0960319Z FBarrier Max Size: 32 2025-12-04T08:54:09.0960454Z *** Done *** 2025-12-04T08:54:09.0970320Z ##[group]Run ngpu=$(rocminfo | grep -c -E 'Name:.*\sgfx') 2025-12-04T08:54:09.0970505Z ngpu=$(rocminfo | grep -c -E 'Name:.*\sgfx') 2025-12-04T08:54:09.0970779Z msg="Please file an issue on pytorch/pytorch reporting the faulty runner. Include a link to the runner logs so the runner can be identified" 2025-12-04T08:54:09.0971043Z if [[ $ngpu -eq 0 ]]; then 2025-12-04T08:54:09.0971190Z  echo "Error: Failed to detect any GPUs on the runner" 2025-12-04T08:54:09.0971332Z  echo "$msg" 2025-12-04T08:54:09.0971439Z  exit 1 2025-12-04T08:54:09.0971539Z fi 2025-12-04T08:54:09.0974299Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:09.0974449Z env: 2025-12-04T08:54:09.0974539Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:09.0974648Z ##[endgroup] 2025-12-04T08:54:09.1514148Z ##[group]Run pytorch/pytorch/.github/actions/diskspace-cleanup@main 2025-12-04T08:54:09.1514316Z with: 2025-12-04T08:54:09.1514416Z diskspace-cutoff: 70 2025-12-04T08:54:09.1514515Z env: 2025-12-04T08:54:09.1514611Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:09.1514714Z ##[endgroup] 2025-12-04T08:54:09.1533563Z ##[group]Run set -ex 2025-12-04T08:54:09.1533696Z set -ex 2025-12-04T08:54:09.1533885Z diskspace_cutoff=70 2025-12-04T08:54:09.1534035Z docker_root_dir=$(docker info -f '{{.DockerRootDir}}') 2025-12-04T08:54:09.1534192Z if [ ! -d "$docker_root_dir" ]; then 2025-12-04T08:54:09.1534393Z  echo "Docker root directory ($docker_root_dir) does not exist. Skipping disk space check." 2025-12-04T08:54:09.1534581Z  exit 0 2025-12-04T08:54:09.1534673Z fi 2025-12-04T08:54:09.1534839Z diskspace=$(df -H --output=pcent ${docker_root_dir} | sed -n 2p | sed 's/%//' | sed 's/ //') 2025-12-04T08:54:09.1535161Z msg="Please file an issue on pytorch/pytorch reporting the faulty runner. Include a link to the runner logs so the runner can be identified" 2025-12-04T08:54:09.1535444Z if [[ "$diskspace" -ge "$diskspace_cutoff" ]] ; then 2025-12-04T08:54:09.1535594Z  docker system prune -af 2025-12-04T08:54:09.1535786Z  diskspace_new=$(df -H --output=pcent ${docker_root_dir} | sed -n 2p | sed 's/%//' | sed 's/ //') 2025-12-04T08:54:09.1536074Z  if [[ "$diskspace_new" -gt "$diskspace_cutoff" ]] ; then 2025-12-04T08:54:09.1536239Z  diskspace_cutoff_int=$((diskspace_cutoff + 0)) 2025-12-04T08:54:09.1536391Z  difference=$((100 - diskspace_cutoff_int)) 2025-12-04T08:54:09.1536599Z  echo "Error: Available diskspace is less than $difference percent. Not enough diskspace." 2025-12-04T08:54:09.1536867Z  echo "$msg" 2025-12-04T08:54:09.1536970Z  exit 1 2025-12-04T08:54:09.1537079Z  else 2025-12-04T08:54:09.1537195Z  difference=$((diskspace - diskspace_new)) 2025-12-04T08:54:09.1537346Z  echo "Diskspace saved: $difference percent" 2025-12-04T08:54:09.1537478Z  fi 2025-12-04T08:54:09.1537568Z fi 2025-12-04T08:54:09.1540388Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:09.1540541Z env: 2025-12-04T08:54:09.1540641Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:09.1540745Z ##[endgroup] 2025-12-04T08:54:09.1555230Z + diskspace_cutoff=70 2025-12-04T08:54:09.1557436Z ++ docker info -f '{{.DockerRootDir}}' 2025-12-04T08:54:09.1860312Z + docker_root_dir=/home/runner/docker-data 2025-12-04T08:54:09.1860973Z + '[' '!' -d /home/runner/docker-data ']' 2025-12-04T08:54:09.1866316Z ++ df -H --output=pcent /home/runner/docker-data 2025-12-04T08:54:09.1866491Z ++ sed -n 2p 2025-12-04T08:54:09.1867710Z ++ sed s/%// 2025-12-04T08:54:09.1868160Z ++ sed 's/ //' 2025-12-04T08:54:09.1880709Z + diskspace=' 4' 2025-12-04T08:54:09.1881028Z + msg='Please file an issue on pytorch/pytorch reporting the faulty runner. Include a link to the runner logs so the runner can be identified' 2025-12-04T08:54:09.1881331Z + [[ 4 -ge 70 ]] 2025-12-04T08:54:09.1894545Z ##[group]Run RUNNER_ARTIFACT_DIR="${RUNNER_TEMP}/artifacts" 2025-12-04T08:54:09.1894743Z RUNNER_ARTIFACT_DIR="${RUNNER_TEMP}/artifacts" 2025-12-04T08:54:09.1894907Z rm -rf "${RUNNER_ARTIFACT_DIR}" 2025-12-04T08:54:09.1895048Z mkdir -p "${RUNNER_ARTIFACT_DIR}" 2025-12-04T08:54:09.1895220Z echo "RUNNER_ARTIFACT_DIR=${RUNNER_ARTIFACT_DIR}" >> "${GITHUB_ENV}" 2025-12-04T08:54:09.1895380Z  2025-12-04T08:54:09.1895503Z RUNNER_TEST_RESULTS_DIR="${RUNNER_TEMP}/test-results" 2025-12-04T08:54:09.1895668Z rm -rf "${RUNNER_TEST_RESULTS_DIR}" 2025-12-04T08:54:09.1895803Z mkdir -p "${RUNNER_TEST_RESULTS_DIR}" 2025-12-04T08:54:09.1895988Z echo "RUNNER_TEST_RESULTS_DIR=${RUNNER_TEST_RESULTS_DIR}" >> "${GITHUB_ENV}" 2025-12-04T08:54:09.1896155Z  2025-12-04T08:54:09.1896257Z RUNNER_DOCS_DIR="${RUNNER_TEMP}/docs" 2025-12-04T08:54:09.1896392Z rm -rf "${RUNNER_DOCS_DIR}" 2025-12-04T08:54:09.1896515Z mkdir -p "${RUNNER_DOCS_DIR}" 2025-12-04T08:54:09.1896670Z echo "RUNNER_DOCS_DIR=${RUNNER_DOCS_DIR}" >> "${GITHUB_ENV}" 2025-12-04T08:54:09.1899719Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:09.1899861Z env: 2025-12-04T08:54:09.1899957Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:09.1900063Z ##[endgroup] 2025-12-04T08:54:09.1973838Z ##[group]Run env | grep '^GITHUB' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-12-04T08:54:09.1974095Z env | grep '^GITHUB' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-12-04T08:54:09.1974285Z env | grep '^CI' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-12-04T08:54:09.1978861Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:09.1979011Z env: 2025-12-04T08:54:09.1979110Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:09.1979249Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T08:54:09.1979421Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T08:54:09.1979592Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T08:54:09.1979867Z ##[endgroup] 2025-12-04T08:54:09.2019358Z ##[group]Run # All GPUs are visible to the runner; visibility, if needed, will be set by run_test.py. 2025-12-04T08:54:09.2019645Z # All GPUs are visible to the runner; visibility, if needed, will be set by run_test.py. 2025-12-04T08:54:09.2019850Z # Add render group for container creation. 2025-12-04T08:54:09.2020022Z render_gid=`cat /etc/group | grep render | cut -d: -f3` 2025-12-04T08:54:09.2020221Z # Ensure GPU isolation if pod is part of kubernetes setup with DEVICE_FLAG. 2025-12-04T08:54:09.2020424Z if [ -f "/etc/podinfo/gha-render-devices" ]; then 2025-12-04T08:54:09.2020594Z  DEVICE_FLAG=$(cat /etc/podinfo/gha-render-devices) 2025-12-04T08:54:09.2020737Z else 2025-12-04T08:54:09.2020845Z  DEVICE_FLAG="--device /dev/dri" 2025-12-04T08:54:09.2020962Z fi 2025-12-04T08:54:09.2021147Z # The --group-add daemon and --group-add bin are needed in the Ubuntu 24.04 and Almalinux OSs respectively. 2025-12-04T08:54:09.2021435Z # This is due to the device files (/dev/kfd & /dev/dri) being owned by video group on bare metal. 2025-12-04T08:54:09.2021691Z # This video group ID maps to subgid 1 inside the docker image due to the /etc/subgid entries. 2025-12-04T08:54:09.2021960Z # The group name corresponding to group ID 1 can change depending on the OS, so both are necessary. 2025-12-04T08:54:09.2022404Z echo "GPU_FLAG=--device=/dev/mem --device=/dev/kfd $DEVICE_FLAG --group-add video --group-add $render_gid --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host" >> "${GITHUB_ENV}" 2025-12-04T08:54:09.2025413Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:09.2025555Z env: 2025-12-04T08:54:09.2025648Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:09.2025779Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T08:54:09.2025959Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T08:54:09.2026125Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T08:54:09.2026251Z ##[endgroup] 2025-12-04T08:54:09.2090553Z ##[group]Run aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722 2025-12-04T08:54:09.2090759Z with: 2025-12-04T08:54:09.2090904Z role-to-assume: arn:aws:iam::308535385114:role/gha_workflow_s3_and_ecr_read_only 2025-12-04T08:54:09.2091070Z aws-region: us-east-1 2025-12-04T08:54:09.2091180Z role-duration-seconds: 18000 2025-12-04T08:54:09.2091295Z audience: sts.amazonaws.com 2025-12-04T08:54:09.2091401Z env: 2025-12-04T08:54:09.2091488Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:09.2091617Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T08:54:09.2091789Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T08:54:09.2091947Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T08:54:09.2092481Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T08:54:09.2092854Z ##[endgroup] 2025-12-04T08:54:09.5630971Z Assuming role with OIDC 2025-12-04T08:54:09.9045923Z Authenticated as assumedRoleId AROAUPVRELQNLLCOPFEJR:GitHubActions 2025-12-04T08:54:10.0001723Z ##[group]Run aws-actions/amazon-ecr-login@062b18b96a7aff071d4dc91bc00c4c1a7945b076 2025-12-04T08:54:10.0001932Z with: 2025-12-04T08:54:10.0002043Z mask-password: true 2025-12-04T08:54:10.0002164Z registry-type: private 2025-12-04T08:54:10.0002285Z skip-logout: false 2025-12-04T08:54:10.0002394Z env: 2025-12-04T08:54:10.0002500Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:10.0002649Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T08:54:10.0002832Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T08:54:10.0003152Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T08:54:10.0003568Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T08:54:10.0003959Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T08:54:10.0004087Z AWS_REGION: us-east-1 2025-12-04T08:54:10.0004505Z AWS_ACCESS_KEY_ID: *** 2025-12-04T08:54:10.0004678Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T08:54:10.0007076Z AWS_SESSION_TOKEN: *** 2025-12-04T08:54:10.0007192Z ##[endgroup] 2025-12-04T08:54:10.4049460Z Logging into registry 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:11.0355142Z ##[group]Run env | grep '^GITHUB' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-12-04T08:54:11.0355452Z env | grep '^GITHUB' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-12-04T08:54:11.0355704Z env | grep '^CI' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-12-04T08:54:11.0355969Z env | grep '^RUNNER' >> "${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" 2025-12-04T08:54:11.0361006Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:11.0361190Z env: 2025-12-04T08:54:11.0361318Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:11.0361489Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T08:54:11.0361725Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T08:54:11.0361935Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T08:54:11.0362414Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T08:54:11.0362794Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T08:54:11.0362917Z AWS_REGION: us-east-1 2025-12-04T08:54:11.0363149Z AWS_ACCESS_KEY_ID: *** 2025-12-04T08:54:11.0363316Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T08:54:11.0365605Z AWS_SESSION_TOKEN: *** 2025-12-04T08:54:11.0365720Z ##[endgroup] 2025-12-04T08:54:11.0508401Z ##[group]Run pytorch/test-infra/.github/actions/calculate-docker-image@main 2025-12-04T08:54:11.0508600Z with: 2025-12-04T08:54:11.0508886Z docker-image-name: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:11.0509198Z use-custom-docker-registry: true 2025-12-04T08:54:11.0509336Z docker-build-dir: .ci/docker 2025-12-04T08:54:11.0509465Z docker-build-script: ./build.sh 2025-12-04T08:54:11.0509594Z working-directory: . 2025-12-04T08:54:11.0509747Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:11.0509911Z force-push: false 2025-12-04T08:54:11.0510017Z env: 2025-12-04T08:54:11.0510115Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:11.0510260Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T08:54:11.0510451Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T08:54:11.0510638Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T08:54:11.0511033Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T08:54:11.0511412Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T08:54:11.0511536Z AWS_REGION: us-east-1 2025-12-04T08:54:11.0511774Z AWS_ACCESS_KEY_ID: *** 2025-12-04T08:54:11.0511941Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T08:54:11.0514135Z AWS_SESSION_TOKEN: *** 2025-12-04T08:54:11.0514247Z ##[endgroup] 2025-12-04T08:54:11.0522921Z ##[group]Run set -ex 2025-12-04T08:54:11.0523061Z set -ex 2025-12-04T08:54:11.0523155Z  2025-12-04T08:54:11.0523310Z # If the docker build directory or the build script doesn't exist, the action will 2025-12-04T08:54:11.0523682Z # gracefully return the docker image name as it is. Pulling docker image in Linux 2025-12-04T08:54:11.0523900Z # job could then download the pre-built image as usual 2025-12-04T08:54:11.0524159Z if [[ -d "${DOCKER_BUILD_DIR}" ]] && [[ -f "${DOCKER_BUILD_DIR}/${DOCKER_BUILD_SCRIPT}" ]] && [[ "${USE_CUSTOM_DOCKER_REGISTRY}" == "true" ]]; then 2025-12-04T08:54:11.0524400Z  echo "skip=false" >> "${GITHUB_OUTPUT}" 2025-12-04T08:54:11.0524532Z else 2025-12-04T08:54:11.0524647Z  echo "skip=true" >> "${GITHUB_OUTPUT}" 2025-12-04T08:54:11.0524822Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:54:11.0524980Z  2025-12-04T08:54:11.0525191Z  echo "Not using custom ECR registry. Either it was not requested or there is no Docker build script in the ${REPO_NAME} repo..." 2025-12-04T08:54:11.0525423Z  exit 0 2025-12-04T08:54:11.0525515Z fi 2025-12-04T08:54:11.0525605Z  2025-12-04T08:54:11.0525743Z if [[ "${DOCKER_IMAGE_NAME}" == *"${DOCKER_REGISTRY}/${REPO_NAME}"* ]]; then 2025-12-04T08:54:11.0525970Z  # The docker image name already includes the ECR prefix and tag, so we can just 2025-12-04T08:54:11.0526171Z  # use it as it is, but first let's extract the tag 2025-12-04T08:54:11.0526356Z  DOCKER_TAG=$(echo "${DOCKER_IMAGE_NAME}" | awk -F '[:,]' '{print $2}') 2025-12-04T08:54:11.0526554Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:54:11.0526738Z  echo "docker-image=${DOCKER_IMAGE_NAME}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:54:11.0526943Z else 2025-12-04T08:54:11.0527057Z  if [[ "${DOCKER_IMAGE_NAME}" == *:* ]]; then 2025-12-04T08:54:11.0527210Z  CUSTOM_TAG_PREFIX=${DOCKER_IMAGE_NAME#*:} 2025-12-04T08:54:11.0527369Z  DOCKER_IMAGE_NAME=${DOCKER_IMAGE_NAME%%:*} 2025-12-04T08:54:11.0527503Z  fi 2025-12-04T08:54:11.0527801Z  DOCKER_TAG=${CUSTOM_TAG_PREFIX:+${CUSTOM_TAG_PREFIX}-}$(git rev-parse HEAD:"${DOCKER_BUILD_DIR}") 2025-12-04T08:54:11.0528034Z  echo "docker-tag=${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:54:11.0528278Z  echo "docker-image=${DOCKER_REGISTRY}/${REPO_NAME}/${DOCKER_IMAGE_NAME}:${DOCKER_TAG}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:54:11.0528543Z  echo "custom-tag-prefix=${CUSTOM_TAG_PREFIX}" >> "${GITHUB_OUTPUT}" 2025-12-04T08:54:11.0528710Z fi 2025-12-04T08:54:11.0532781Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:11.0532928Z env: 2025-12-04T08:54:11.0533027Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:11.0533169Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T08:54:11.0533353Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T08:54:11.0533528Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T08:54:11.0533922Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T08:54:11.0534304Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T08:54:11.0534424Z AWS_REGION: us-east-1 2025-12-04T08:54:11.0534569Z AWS_ACCESS_KEY_ID: *** 2025-12-04T08:54:11.0534728Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T08:54:11.0536963Z AWS_SESSION_TOKEN: *** 2025-12-04T08:54:11.0537075Z REPO_NAME: pytorch 2025-12-04T08:54:11.0537361Z DOCKER_IMAGE_NAME: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:11.0537661Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T08:54:11.0537785Z DOCKER_BUILD_SCRIPT: ./build.sh 2025-12-04T08:54:11.0537944Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:11.0538112Z USE_CUSTOM_DOCKER_REGISTRY: true 2025-12-04T08:54:11.0538286Z CUSTOM_TAG_PREFIX: 2025-12-04T08:54:11.0538398Z ##[endgroup] 2025-12-04T08:54:11.0550937Z + [[ -d .ci/docker ]] 2025-12-04T08:54:11.0551063Z + [[ -f .ci/docker/./build.sh ]] 2025-12-04T08:54:11.0551191Z + [[ true == \t\r\u\e ]] 2025-12-04T08:54:11.0551304Z + echo skip=false 2025-12-04T08:54:11.0551874Z + [[ 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a == *\3\0\8\5\3\5\3\8\5\1\1\4\.\d\k\r\.\e\c\r\.\u\s\-\e\a\s\t\-\1\.\a\m\a\z\o\n\a\w\s\.\c\o\m\/\p\y\t\o\r\c\h* ]] 2025-12-04T08:54:11.0557531Z ++ echo 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:11.0557959Z ++ awk -F '[:,]' '{print $2}' 2025-12-04T08:54:11.0569555Z + DOCKER_TAG=pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:11.0569981Z + echo docker-tag=pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:11.0570541Z + echo docker-image=308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:11.0599849Z ##[group]Run set +e 2025-12-04T08:54:11.0600029Z set +e 2025-12-04T08:54:11.0600147Z set -x 2025-12-04T08:54:11.0600261Z  2025-12-04T08:54:11.0600372Z login() { 2025-12-04T08:54:11.0600613Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T08:54:11.0600851Z } 2025-12-04T08:54:11.0600962Z  2025-12-04T08:54:11.0601073Z retry () { 2025-12-04T08:54:11.0601215Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T08:54:11.0601367Z } 2025-12-04T08:54:11.0601472Z  2025-12-04T08:54:11.0601587Z retry login "${DOCKER_REGISTRY}" 2025-12-04T08:54:11.0601757Z  2025-12-04T08:54:11.0601876Z START_TIME=$(date +%s) 2025-12-04T08:54:11.0602049Z # Wait up to 120 minutes 2025-12-04T08:54:11.0602403Z while [[ $(( $(date +%s) - 7200 )) -lt $START_TIME ]]; do 2025-12-04T08:54:11.0602637Z  # Check if image already exists, if it does then skip building it 2025-12-04T08:54:11.0602868Z  if docker manifest inspect "${DOCKER_IMAGE}"; then 2025-12-04T08:54:11.0603044Z  exit 0 2025-12-04T08:54:11.0603168Z  fi 2025-12-04T08:54:11.0603283Z  2025-12-04T08:54:11.0603467Z  # NB: This flag is used by Docker build workflow to push the image to ECR, so we can 2025-12-04T08:54:11.0603767Z  # use this to differentiate between the Docker build and regular build jobs. For the 2025-12-04T08:54:11.0604064Z  # latter, it will wait for the Docker images to become available before continuing 2025-12-04T08:54:11.0604309Z  if [ "${DOCKER_PUSH:-false}" == "true" ]; then 2025-12-04T08:54:11.0604505Z  # It's a Docker build job, let's build the image 2025-12-04T08:54:11.0604679Z  break 2025-12-04T08:54:11.0604806Z  else 2025-12-04T08:54:11.0604974Z  # It's a regular build job, wait for the image to become available 2025-12-04T08:54:11.0605172Z  sleep 300 2025-12-04T08:54:11.0605299Z  fi 2025-12-04T08:54:11.0605414Z done 2025-12-04T08:54:11.0605526Z  2025-12-04T08:54:11.0605688Z # NB: This part requires a full checkout. Otherwise, the merge base will 2025-12-04T08:54:11.0605904Z # be empty. The default action would be to continue rebuild the image 2025-12-04T08:54:11.0606099Z if [[ "$BASE_REVISION" = "$(git rev-parse HEAD)" ]]; then 2025-12-04T08:54:11.0606280Z  # if we're on the base branch then use the parent commit 2025-12-04T08:54:11.0606440Z  MERGE_BASE=$(git rev-parse HEAD~) 2025-12-04T08:54:11.0606567Z else 2025-12-04T08:54:11.0606698Z  # otherwise we're on a PR, so use the most recent base commit 2025-12-04T08:54:11.0607212Z  MERGE_BASE=$(git merge-base HEAD "$BASE_REVISION") 2025-12-04T08:54:11.0607351Z fi 2025-12-04T08:54:11.0607439Z  2025-12-04T08:54:11.0607536Z if [[ -z "${MERGE_BASE}" ]]; then 2025-12-04T08:54:11.0607681Z  echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T08:54:11.0607809Z  2025-12-04T08:54:11.0607986Z  echo "Finding merge base only works with full checkout, please set fetch-depth to 0, continuing ..." 2025-12-04T08:54:11.0608186Z  exit 0 2025-12-04T08:54:11.0608278Z fi 2025-12-04T08:54:11.0608364Z  2025-12-04T08:54:11.0608488Z if ! git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}"; then 2025-12-04T08:54:11.0608739Z  echo "Directory '${DOCKER_BUILD_DIR}' not found in commit $MERGE_BASE, you should rebase onto a more recent commit" 2025-12-04T08:54:11.0608952Z  exit 1 2025-12-04T08:54:11.0609051Z fi 2025-12-04T08:54:11.0609138Z  2025-12-04T08:54:11.0609284Z PREVIOUS_DOCKER_TAG=$(git rev-parse "${MERGE_BASE}:${DOCKER_BUILD_DIR}") 2025-12-04T08:54:11.0609531Z # If no image exists but the hash is the same as the previous hash then we should error out here 2025-12-04T08:54:11.0609748Z if [[ "${PREVIOUS_DOCKER_TAG}" == "${DOCKER_TAG}" ]]; then 2025-12-04T08:54:11.0609999Z  echo "WARNING: Something has gone wrong and the previous image isn't available for the merge-base of your branch" 2025-12-04T08:54:11.0610279Z  echo " Will re-build docker image to store in local cache, TTS may be longer" 2025-12-04T08:54:11.0610450Z fi 2025-12-04T08:54:11.0610537Z  2025-12-04T08:54:11.0610648Z echo "rebuild=true" >> "${GITHUB_OUTPUT}" 2025-12-04T08:54:11.0614842Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:11.0614992Z env: 2025-12-04T08:54:11.0615093Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:11.0615229Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T08:54:11.0615455Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T08:54:11.0615623Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T08:54:11.0616009Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T08:54:11.0616383Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T08:54:11.0616500Z AWS_REGION: us-east-1 2025-12-04T08:54:11.0616707Z AWS_ACCESS_KEY_ID: *** 2025-12-04T08:54:11.0616911Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T08:54:11.0619113Z AWS_SESSION_TOKEN: *** 2025-12-04T08:54:11.0619224Z DOCKER_BUILD_DIR: .ci/docker 2025-12-04T08:54:11.0619367Z BASE_REVISION: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T08:54:11.0619687Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:11.0620056Z DOCKER_TAG: pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:11.0620293Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:11.0620448Z DOCKER_PUSH: 2025-12-04T08:54:11.0620551Z ##[endgroup] 2025-12-04T08:54:11.0636957Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:11.0637147Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:11.0639804Z + aws ecr get-login-password --region us-east-1 2025-12-04T08:54:11.0640234Z /home/runner/_work/_temp/e6cefd4a-99be-4852-b8e3-d64c2f5dbbeb.sh: line 5: aws: command not found 2025-12-04T08:54:11.0640738Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:11.0738348Z Error: Cannot perform an interactive login from a non TTY device 2025-12-04T08:54:11.0746957Z + sleep 1 2025-12-04T08:54:12.0756538Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:12.0759926Z + aws ecr get-login-password --region us-east-1 2025-12-04T08:54:12.0760457Z /home/runner/_work/_temp/e6cefd4a-99be-4852-b8e3-d64c2f5dbbeb.sh: line 5: aws: command not found 2025-12-04T08:54:12.0761362Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:12.0860233Z Error: Cannot perform an interactive login from a non TTY device 2025-12-04T08:54:12.0870873Z + sleep 2 2025-12-04T08:54:14.0882420Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:14.0886143Z + aws ecr get-login-password --region us-east-1 2025-12-04T08:54:14.0886719Z /home/runner/_work/_temp/e6cefd4a-99be-4852-b8e3-d64c2f5dbbeb.sh: line 5: aws: command not found 2025-12-04T08:54:14.0887416Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:14.0974386Z Error: Cannot perform an interactive login from a non TTY device 2025-12-04T08:54:14.0986453Z ++ date +%s 2025-12-04T08:54:14.0993200Z + START_TIME=1764838454 2025-12-04T08:54:14.0995973Z ++ date +%s 2025-12-04T08:54:14.1002125Z + [[ 1764831254 -lt 1764838454 ]] 2025-12-04T08:54:14.1002730Z + docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:15.4574504Z { 2025-12-04T08:54:15.4574916Z "schemaVersion": 2, 2025-12-04T08:54:15.4575396Z "mediaType": "application/vnd.docker.distribution.manifest.v2+json", 2025-12-04T08:54:15.4575818Z "config": { 2025-12-04T08:54:15.4576150Z "mediaType": "application/vnd.docker.container.image.v1+json", 2025-12-04T08:54:15.4576517Z "size": 30520, 2025-12-04T08:54:15.4576967Z "digest": "sha256:45252333063339f104d56e41f20304e9511ab21c7768e8d156b95ddf24a9dbe5" 2025-12-04T08:54:15.4577371Z }, 2025-12-04T08:54:15.4577570Z "layers": [ 2025-12-04T08:54:15.4577771Z { 2025-12-04T08:54:15.4578108Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4578490Z "size": 30447951, 2025-12-04T08:54:15.4579495Z "digest": "sha256:63e5bc7682b85ae57a1221210f64d62e7a90b0a30f19af4ca734b8242ae49d63" 2025-12-04T08:54:15.4579913Z }, 2025-12-04T08:54:15.4580107Z { 2025-12-04T08:54:15.4580415Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4580780Z "size": 1554, 2025-12-04T08:54:15.4581148Z "digest": "sha256:835841cca3b7e1464290cdb78e48773e03583413fbed852c3cc5165a392ea44d" 2025-12-04T08:54:15.4581558Z }, 2025-12-04T08:54:15.4581742Z { 2025-12-04T08:54:15.4582046Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4582414Z "size": 313275691, 2025-12-04T08:54:15.4582801Z "digest": "sha256:aac69780afc8611a5f94a235792d39ae055249c8319ef43b78675998a9b2f825" 2025-12-04T08:54:15.4583208Z }, 2025-12-04T08:54:15.4583394Z { 2025-12-04T08:54:15.4583694Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4584064Z "size": 704, 2025-12-04T08:54:15.4584389Z "digest": "sha256:029495b23122c840ca0e52d487afa8d2c4dbf1991cd7f204ec3e434dcf947bf4" 2025-12-04T08:54:15.4584700Z }, 2025-12-04T08:54:15.4584838Z { 2025-12-04T08:54:15.4585069Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4599264Z "size": 1218, 2025-12-04T08:54:15.4599568Z "digest": "sha256:d0fb85b008332051a3f7c052721ef68bde404b46c23fa43ad040373bd367826c" 2025-12-04T08:54:15.4599978Z }, 2025-12-04T08:54:15.4600176Z { 2025-12-04T08:54:15.4600395Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4600616Z "size": 484, 2025-12-04T08:54:15.4600844Z "digest": "sha256:59b63930883363c7d2aaab27cc61555d9f3e119dc18247a8624c98ebdaa354a5" 2025-12-04T08:54:15.4601075Z }, 2025-12-04T08:54:15.4601199Z { 2025-12-04T08:54:15.4601375Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4601771Z "size": 110363202, 2025-12-04T08:54:15.4602003Z "digest": "sha256:dc112c89d57aa1e85082e40a56e5bc743d64f834ae2f98afe91f60c248354d38" 2025-12-04T08:54:15.4602242Z }, 2025-12-04T08:54:15.4602355Z { 2025-12-04T08:54:15.4602543Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4602753Z "size": 4436, 2025-12-04T08:54:15.4602970Z "digest": "sha256:522eab2402e5001810155ef7eb56940b7c01a4fef62ac588886981c3b8ee8e1e" 2025-12-04T08:54:15.4603207Z }, 2025-12-04T08:54:15.4603311Z { 2025-12-04T08:54:15.4603481Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4603683Z "size": 1755, 2025-12-04T08:54:15.4603890Z "digest": "sha256:2b5a11b41761d8ea3b829e4772e4064cb6c4e4989126af324d0057661e4493a1" 2025-12-04T08:54:15.4604121Z }, 2025-12-04T08:54:15.4604246Z { 2025-12-04T08:54:15.4604421Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4604588Z "size": 724, 2025-12-04T08:54:15.4604758Z "digest": "sha256:9681563a88ff9e62494a2740e537440d3df978d466c9478d6a941fae8b57b084" 2025-12-04T08:54:15.4604940Z }, 2025-12-04T08:54:15.4605021Z { 2025-12-04T08:54:15.4605157Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4605322Z "size": 3185588166, 2025-12-04T08:54:15.4605494Z "digest": "sha256:73e33534e9eb94cf29418d65944168962b65fe21f55e9b8bad18c76e9b3a37b8" 2025-12-04T08:54:15.4605675Z }, 2025-12-04T08:54:15.4605760Z { 2025-12-04T08:54:15.4605892Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4606054Z "size": 396, 2025-12-04T08:54:15.4606228Z "digest": "sha256:5bfdaeb5578d6ffcd7db29c48303cbceb13c591210feaa216a8daa7a6d445b4b" 2025-12-04T08:54:15.4606415Z }, 2025-12-04T08:54:15.4606500Z { 2025-12-04T08:54:15.4606633Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4606857Z "size": 236863, 2025-12-04T08:54:15.4607034Z "digest": "sha256:c07d27e4d3a5ba4ad5325bb785b2e4f058fe5e10ec1aeeb413a1e152b073f203" 2025-12-04T08:54:15.4607229Z }, 2025-12-04T08:54:15.4607393Z { 2025-12-04T08:54:15.4607528Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4607691Z "size": 787, 2025-12-04T08:54:15.4607862Z "digest": "sha256:b21856d1bf420da6fa8ec7331b82ab355d4f4178644e7d3a3d3d0fbc3610109a" 2025-12-04T08:54:15.4608048Z }, 2025-12-04T08:54:15.4608133Z { 2025-12-04T08:54:15.4608269Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4608431Z "size": 106, 2025-12-04T08:54:15.4608600Z "digest": "sha256:cb19d84867e4063f55db9459c28c50a2abc37c06d3c1ca82ba95fa8427cc438a" 2025-12-04T08:54:15.4608784Z }, 2025-12-04T08:54:15.4608868Z { 2025-12-04T08:54:15.4609004Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4609168Z "size": 1496, 2025-12-04T08:54:15.4609336Z "digest": "sha256:8165374f8dccf88a7791a5d31afbe29e4d4542b4f1cf1904945e07f9af6bf8ba" 2025-12-04T08:54:15.4609525Z }, 2025-12-04T08:54:15.4609611Z { 2025-12-04T08:54:15.4609748Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4609914Z "size": 458789560, 2025-12-04T08:54:15.4610089Z "digest": "sha256:1aecc77354ceba59ec6f0d37a558f2dbb6d5c0854553ee8505ac8707b422da6d" 2025-12-04T08:54:15.4610280Z }, 2025-12-04T08:54:15.4610366Z { 2025-12-04T08:54:15.4610499Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4610661Z "size": 164, 2025-12-04T08:54:15.4610836Z "digest": "sha256:465d3fd643aa2ea0ad07335cda66f12f1d7e5e800c4e9385ec466bc8a1ceabda" 2025-12-04T08:54:15.4611024Z }, 2025-12-04T08:54:15.4611114Z { 2025-12-04T08:54:15.4611254Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4611422Z "size": 104, 2025-12-04T08:54:15.4611598Z "digest": "sha256:6c503e779d6f41ca7f51309875df2b725c171926aece7009c4b8a64d1ba3f58e" 2025-12-04T08:54:15.4611837Z }, 2025-12-04T08:54:15.4611930Z { 2025-12-04T08:54:15.4612072Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4612240Z "size": 724, 2025-12-04T08:54:15.4612405Z "digest": "sha256:9681563a88ff9e62494a2740e537440d3df978d466c9478d6a941fae8b57b084" 2025-12-04T08:54:15.4612591Z }, 2025-12-04T08:54:15.4612678Z { 2025-12-04T08:54:15.4612821Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4612993Z "size": 196, 2025-12-04T08:54:15.4613167Z "digest": "sha256:f7e9a021f0ee3d11a50dcb96378af8103a21f6c3c142f54529207648f3ed00b2" 2025-12-04T08:54:15.4613353Z }, 2025-12-04T08:54:15.4613435Z { 2025-12-04T08:54:15.4613579Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4613739Z "size": 2583, 2025-12-04T08:54:15.4613903Z "digest": "sha256:8e023b349080fb11ee55491bc9b842b30e9e3a90246d05b303a73dc62038caf2" 2025-12-04T08:54:15.4614080Z }, 2025-12-04T08:54:15.4614172Z { 2025-12-04T08:54:15.4614299Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4614458Z "size": 7577171420, 2025-12-04T08:54:15.4614622Z "digest": "sha256:8188df80e595a3dbcf84623c6a58a655269898cbb60029435f136d7f9d34ccaa" 2025-12-04T08:54:15.4614797Z }, 2025-12-04T08:54:15.4614878Z { 2025-12-04T08:54:15.4614987Z + exit 0 2025-12-04T08:54:15.4615122Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4615276Z "size": 135, 2025-12-04T08:54:15.4615438Z "digest": "sha256:3c2c2f8c74bfa16c4bf9a832c97bbb1d55205b2b4a2cead02cf74301ca1001fb" 2025-12-04T08:54:15.4615618Z }, 2025-12-04T08:54:15.4615698Z { 2025-12-04T08:54:15.4615827Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4615981Z "size": 104, 2025-12-04T08:54:15.4616145Z "digest": "sha256:2aa7784fbe3300f8bbfb6bb51cff3b01fd091e829c2bc7ab9e25261a0dd9b3bd" 2025-12-04T08:54:15.4616324Z }, 2025-12-04T08:54:15.4616413Z { 2025-12-04T08:54:15.4616543Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4616734Z "size": 612, 2025-12-04T08:54:15.4616932Z "digest": "sha256:2b3b5215d3ebe8789f0444457bfd5a6e218289b64aa07653ac3d03ddda5e6708" 2025-12-04T08:54:15.4617107Z }, 2025-12-04T08:54:15.4617188Z { 2025-12-04T08:54:15.4617318Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4617475Z "size": 838191945, 2025-12-04T08:54:15.4617645Z "digest": "sha256:99b1f1ea3e857834cebd01763d90fbd700aeb9c2d2ef23eda2cfff5652c9708b" 2025-12-04T08:54:15.4617826Z }, 2025-12-04T08:54:15.4617906Z { 2025-12-04T08:54:15.4618034Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4618189Z "size": 111, 2025-12-04T08:54:15.4618350Z "digest": "sha256:18d6daba0a5768a37ad106b57974f6b7efd35c43a87c246bcd3f43fea88f2d2b" 2025-12-04T08:54:15.4618529Z }, 2025-12-04T08:54:15.4618610Z { 2025-12-04T08:54:15.4618743Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4618899Z "size": 1555, 2025-12-04T08:54:15.4619065Z "digest": "sha256:5277f2a503ebd17ba9d9b86cc9bac86265504adeb449c0647616ddaacd3cbc41" 2025-12-04T08:54:15.4619241Z }, 2025-12-04T08:54:15.4619321Z { 2025-12-04T08:54:15.4619450Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4619606Z "size": 107, 2025-12-04T08:54:15.4619766Z "digest": "sha256:3198a9717aace920fd5de085319adf75091af05fc4318ce4b16a8a5b0e8d449e" 2025-12-04T08:54:15.4619942Z }, 2025-12-04T08:54:15.4620024Z { 2025-12-04T08:54:15.4620155Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4620309Z "size": 166, 2025-12-04T08:54:15.4620466Z "digest": "sha256:99a4918e5808277879449e97ccd7190db6b9aa2d742b57a3b831ce0198522bdd" 2025-12-04T08:54:15.4620637Z }, 2025-12-04T08:54:15.4620717Z { 2025-12-04T08:54:15.4620844Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4621046Z "size": 3526081, 2025-12-04T08:54:15.4621212Z "digest": "sha256:15bb11dfc6acc3537d527d6771c8e711e5605e99f82ec41e805d4600b8a97516" 2025-12-04T08:54:15.4621391Z }, 2025-12-04T08:54:15.4621470Z { 2025-12-04T08:54:15.4621599Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4621753Z "size": 107, 2025-12-04T08:54:15.4621916Z "digest": "sha256:bd87c8766e90e33db17514558ac591cc3f4149afd7abeaef4dd5770bbfa14210" 2025-12-04T08:54:15.4622094Z }, 2025-12-04T08:54:15.4622173Z { 2025-12-04T08:54:15.4622301Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4622456Z "size": 829, 2025-12-04T08:54:15.4622615Z "digest": "sha256:1969e15d0c13874ea5883ed829235a19ef6dc21c8aa6172032b78a8ffa6ff262" 2025-12-04T08:54:15.4622789Z }, 2025-12-04T08:54:15.4622869Z { 2025-12-04T08:54:15.4622997Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4623159Z "size": 26973054, 2025-12-04T08:54:15.4623329Z "digest": "sha256:24a03847d382b73c11969f8f73916a6bedf5ccea12f6f4290b3880f29ceda32a" 2025-12-04T08:54:15.4623504Z }, 2025-12-04T08:54:15.4623585Z { 2025-12-04T08:54:15.4623714Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4623869Z "size": 104, 2025-12-04T08:54:15.4624030Z "digest": "sha256:816e2e34e01839a35d624dbf4bd9ac9bea4c975104af47a0e6b6b6dee6c6f98d" 2025-12-04T08:54:15.4624207Z }, 2025-12-04T08:54:15.4624286Z { 2025-12-04T08:54:15.4624414Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4624569Z "size": 424, 2025-12-04T08:54:15.4624728Z "digest": "sha256:b168858b85373f8ddca549d79267a06de4fa945d04bf791c55c9ddc93957fa3c" 2025-12-04T08:54:15.4624901Z }, 2025-12-04T08:54:15.4624980Z { 2025-12-04T08:54:15.4625108Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4625265Z "size": 19309386, 2025-12-04T08:54:15.4625441Z "digest": "sha256:6b8d5ff02e267e38322afbb8a58ed63ce9d75b10e9e73255e6affcbc6b6539bf" 2025-12-04T08:54:15.4625664Z }, 2025-12-04T08:54:15.4625745Z { 2025-12-04T08:54:15.4625875Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4626028Z "size": 826, 2025-12-04T08:54:15.4626188Z "digest": "sha256:4e3b10a5dd6aed29f238d604925e2a4f873141c1087c8dd4fdde5c61e7560893" 2025-12-04T08:54:15.4626363Z }, 2025-12-04T08:54:15.4626446Z { 2025-12-04T08:54:15.4626578Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4626733Z "size": 724, 2025-12-04T08:54:15.4626923Z "digest": "sha256:9681563a88ff9e62494a2740e537440d3df978d466c9478d6a941fae8b57b084" 2025-12-04T08:54:15.4627095Z }, 2025-12-04T08:54:15.4627173Z { 2025-12-04T08:54:15.4627302Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4627455Z "size": 149, 2025-12-04T08:54:15.4627614Z "digest": "sha256:3092fab73b59190b9facfc49bf18f58612172bc2fd68dfa339a1118632616939" 2025-12-04T08:54:15.4627793Z }, 2025-12-04T08:54:15.4627878Z { 2025-12-04T08:54:15.4628007Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4628163Z "size": 136, 2025-12-04T08:54:15.4628327Z "digest": "sha256:20020dd28a15ba092fcbfe906ee39cdddfcc9d0b7eb42fdd6f4c08a984fa9c00" 2025-12-04T08:54:15.4628506Z }, 2025-12-04T08:54:15.4628588Z { 2025-12-04T08:54:15.4628719Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4628874Z "size": 140, 2025-12-04T08:54:15.4629036Z "digest": "sha256:ae5280ce969dcff08c091e9a5f7641f13561b2b0ee44d78b7c3f81d8fe8e6d32" 2025-12-04T08:54:15.4629215Z }, 2025-12-04T08:54:15.4629297Z { 2025-12-04T08:54:15.4629424Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4629580Z "size": 32, 2025-12-04T08:54:15.4629750Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:54:15.4630010Z }, 2025-12-04T08:54:15.4630102Z { 2025-12-04T08:54:15.4630242Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4630407Z "size": 222, 2025-12-04T08:54:15.4630578Z "digest": "sha256:fe17d9eb0fd26d3af4c724bf570d833978b131cedb7dc17a800aa388a246b3cd" 2025-12-04T08:54:15.4630760Z }, 2025-12-04T08:54:15.4630849Z { 2025-12-04T08:54:15.4630988Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4631151Z "size": 346, 2025-12-04T08:54:15.4631318Z "digest": "sha256:a51e0dab2d596e6563483f27c12660007160847d177ba4c31812a8f44ada5754" 2025-12-04T08:54:15.4631498Z }, 2025-12-04T08:54:15.4631584Z { 2025-12-04T08:54:15.4631722Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4631885Z "size": 88300, 2025-12-04T08:54:15.4632061Z "digest": "sha256:6eb176cefd72d37ecbcdf074289a8f1de732d8816cc695ece7e4709d098094d6" 2025-12-04T08:54:15.4632254Z }, 2025-12-04T08:54:15.4632342Z { 2025-12-04T08:54:15.4632482Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4632647Z "size": 106, 2025-12-04T08:54:15.4632815Z "digest": "sha256:e7b8cf2e8d5a4c56db9726ce62c1176032408b3b1c25a000592361cb4245e2b5" 2025-12-04T08:54:15.4632999Z }, 2025-12-04T08:54:15.4633092Z { 2025-12-04T08:54:15.4633232Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4633397Z "size": 1671, 2025-12-04T08:54:15.4633572Z "digest": "sha256:ef3a5060abce88884bc8bd815aa41c46427f34eeb132fe0ddd85a3f86e6dc83d" 2025-12-04T08:54:15.4633762Z }, 2025-12-04T08:54:15.4633848Z { 2025-12-04T08:54:15.4633986Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4634151Z "size": 724, 2025-12-04T08:54:15.4634315Z "digest": "sha256:9681563a88ff9e62494a2740e537440d3df978d466c9478d6a941fae8b57b084" 2025-12-04T08:54:15.4634494Z }, 2025-12-04T08:54:15.4634584Z { 2025-12-04T08:54:15.4634725Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4634936Z "size": 138, 2025-12-04T08:54:15.4635111Z "digest": "sha256:a6f4ec14b42b8f0a83d20aa6a985ddb6a1bf64e0ed3d44afd3484b87d4ed5ad3" 2025-12-04T08:54:15.4635299Z }, 2025-12-04T08:54:15.4635386Z { 2025-12-04T08:54:15.4635527Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4635692Z "size": 119, 2025-12-04T08:54:15.4635865Z "digest": "sha256:7e5a0c956cfbd6f8074fbfd3b1d416e6635d632835ec00c8dd4c015a21da19b4" 2025-12-04T08:54:15.4636053Z }, 2025-12-04T08:54:15.4636145Z { 2025-12-04T08:54:15.4636284Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4636455Z "size": 6238423049, 2025-12-04T08:54:15.4636636Z "digest": "sha256:b4f78730cfe76ce091b78b2e2e3d52be03f1097b3e4c3de5bd79f8d13a853132" 2025-12-04T08:54:15.4636863Z }, 2025-12-04T08:54:15.4636955Z { 2025-12-04T08:54:15.4637097Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4637269Z "size": 174, 2025-12-04T08:54:15.4637439Z "digest": "sha256:081028f24389b112683689fd362e8c0d6f358082710e72feab91cea6383feb4d" 2025-12-04T08:54:15.4637623Z }, 2025-12-04T08:54:15.4637713Z { 2025-12-04T08:54:15.4637854Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4638020Z "size": 1896, 2025-12-04T08:54:15.4638196Z "digest": "sha256:a534dcf4b9a9e5fabed742c8a8fc43c9cfe7346ea88ab3c177c3b14fd3afe00a" 2025-12-04T08:54:15.4638385Z }, 2025-12-04T08:54:15.4638478Z { 2025-12-04T08:54:15.4638617Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4638786Z "size": 197577597, 2025-12-04T08:54:15.4638962Z "digest": "sha256:2e77500302cc13224427e1d74e471bd79d5109ba6a5099a83df1d10b786f71ba" 2025-12-04T08:54:15.4639144Z }, 2025-12-04T08:54:15.4639229Z { 2025-12-04T08:54:15.4639369Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4639583Z "size": 304, 2025-12-04T08:54:15.4639760Z "digest": "sha256:bc08246bb4ba18c3ec5bc69e16b6b4e929c5bd0f3fae10eeb0b1a622a63d6fa2" 2025-12-04T08:54:15.4639950Z }, 2025-12-04T08:54:15.4640039Z { 2025-12-04T08:54:15.4640177Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4640342Z "size": 32, 2025-12-04T08:54:15.4640520Z "digest": "sha256:4f4fb700ef54461cfa02571ae0db9a0dc1e0cdb5577484a6d75e68dc38e8acc1" 2025-12-04T08:54:15.4640708Z }, 2025-12-04T08:54:15.4640796Z { 2025-12-04T08:54:15.4640935Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4641099Z "size": 106, 2025-12-04T08:54:15.4641269Z "digest": "sha256:ff0c473ca120ebdcaa2ba10b3274e82032edd5196019e76d4e7584553704ae81" 2025-12-04T08:54:15.4641454Z }, 2025-12-04T08:54:15.4641540Z { 2025-12-04T08:54:15.4641673Z "mediaType": "application/vnd.docker.image.rootfs.diff.tar.gzip", 2025-12-04T08:54:15.4641843Z "size": 54145662, 2025-12-04T08:54:15.4642022Z "digest": "sha256:6bbc14b250efb3cdaad12c91573c6bb9129ad3e3432f0ed1a7eaebc9958d162f" 2025-12-04T08:54:15.4642213Z } 2025-12-04T08:54:15.4642303Z ] 2025-12-04T08:54:15.4642395Z } 2025-12-04T08:54:15.4660746Z ##[group]Run set -eux 2025-12-04T08:54:15.4660883Z set -eux 2025-12-04T08:54:15.4661057Z # It's ok if this steps fails, it would then be an anonymous user like what we used to have 2025-12-04T08:54:15.4661487Z aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token | jq --raw-output '.SecretString' | jq -r .docker_hub_readonly_token | docker login --username pytorchbot --password-stdin || true 2025-12-04T08:54:15.4666464Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:15.4666629Z env: 2025-12-04T08:54:15.4666736Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:15.4667115Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T08:54:15.4667299Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T08:54:15.4667485Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T08:54:15.4667965Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T08:54:15.4668351Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T08:54:15.4668481Z AWS_REGION: us-east-1 2025-12-04T08:54:15.4668694Z AWS_ACCESS_KEY_ID: *** 2025-12-04T08:54:15.4668861Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T08:54:15.4671118Z AWS_SESSION_TOKEN: *** 2025-12-04T08:54:15.4671235Z ##[endgroup] 2025-12-04T08:54:15.4700254Z + aws secretsmanager get-secret-value --secret-id docker_hub_readonly_token 2025-12-04T08:54:15.4700688Z + jq --raw-output .SecretString 2025-12-04T08:54:15.4701118Z /home/runner/_work/_temp/58d9b1aa-0979-4d4d-9dbb-3c471bb7e254.sh: line 3: aws: command not found 2025-12-04T08:54:15.4701591Z + jq -r .docker_hub_readonly_token 2025-12-04T08:54:15.4701906Z + docker login --username pytorchbot --password-stdin 2025-12-04T08:54:15.4791547Z Error: Cannot perform an interactive login from a non TTY device 2025-12-04T08:54:15.4798111Z + true 2025-12-04T08:54:15.4860034Z ##[group]Run pytorch/test-infra/.github/actions/pull-docker-image@main 2025-12-04T08:54:15.4860214Z with: 2025-12-04T08:54:15.4860486Z docker-image: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:15.4860814Z docker-registry: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:15.4860967Z env: 2025-12-04T08:54:15.4861060Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:15.4861198Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T08:54:15.4861375Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T08:54:15.4861540Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T08:54:15.4862061Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T08:54:15.4862448Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T08:54:15.4862564Z AWS_REGION: us-east-1 2025-12-04T08:54:15.4862754Z AWS_ACCESS_KEY_ID: *** 2025-12-04T08:54:15.4862907Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T08:54:15.4865141Z AWS_SESSION_TOKEN: *** 2025-12-04T08:54:15.4865246Z ##[endgroup] 2025-12-04T08:54:15.4872052Z ##[group]Run set -x 2025-12-04T08:54:15.4872172Z set -x 2025-12-04T08:54:15.4872264Z set +e 2025-12-04T08:54:15.4872355Z  2025-12-04T08:54:15.4872443Z login() { 2025-12-04T08:54:15.4872630Z  aws ecr get-login-password --region us-east-1 | docker login -u AWS --password-stdin "$1" 2025-12-04T08:54:15.4872823Z } 2025-12-04T08:54:15.4872909Z  2025-12-04T08:54:15.4873003Z retry () { 2025-12-04T08:54:15.4873116Z  $* || (sleep 1 && $*) || (sleep 2 && $*) 2025-12-04T08:54:15.4873240Z } 2025-12-04T08:54:15.4873324Z  2025-12-04T08:54:15.4873421Z retry login "${DOCKER_REGISTRY}" 2025-12-04T08:54:15.4873537Z  2025-12-04T08:54:15.4873723Z IMAGE_SIZE=$(docker manifest inspect "${DOCKER_IMAGE}" | jq '[.layers[].size, .config.size] | add / 1024 / 1024') 2025-12-04T08:54:15.4873966Z echo "Compressed size of image in MB: ${IMAGE_SIZE}" 2025-12-04T08:54:15.4874109Z  2025-12-04T08:54:15.4874192Z set -e 2025-12-04T08:54:15.4874329Z # ignore output since only exit code is used for conditional 2025-12-04T08:54:15.4874516Z # only pull docker image if it's not available locally 2025-12-04T08:54:15.4874720Z if ! docker inspect --type=image "${DOCKER_IMAGE}" >/dev/null 2>/dev/null; then 2025-12-04T08:54:15.4874907Z  retry docker pull "${DOCKER_IMAGE}" 2025-12-04T08:54:15.4875037Z fi 2025-12-04T08:54:15.4879418Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T08:54:15.4879563Z env: 2025-12-04T08:54:15.4879655Z GIT_DEFAULT_BRANCH: main 2025-12-04T08:54:15.4879792Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T08:54:15.4879966Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T08:54:15.4880129Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T08:54:15.4880511Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T08:54:15.4880879Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T08:54:15.4880994Z AWS_REGION: us-east-1 2025-12-04T08:54:15.4881132Z AWS_ACCESS_KEY_ID: *** 2025-12-04T08:54:15.4881282Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T08:54:15.4883484Z AWS_SESSION_TOKEN: *** 2025-12-04T08:54:15.4883759Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:15.4884167Z DOCKER_REGISTRY: 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:15.4884319Z ##[endgroup] 2025-12-04T08:54:15.4899361Z + set +e 2025-12-04T08:54:15.4899502Z + retry login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:15.4899672Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:15.4902248Z + aws ecr get-login-password --region us-east-1 2025-12-04T08:54:15.4902563Z /home/runner/_work/_temp/f8be5534-de81-4bb1-a4ee-7faece589516.sh: line 5: aws: command not found 2025-12-04T08:54:15.4903310Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:15.4988045Z Error: Cannot perform an interactive login from a non TTY device 2025-12-04T08:54:15.4995558Z + sleep 1 2025-12-04T08:54:16.5004822Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:16.5007426Z + aws ecr get-login-password --region us-east-1 2025-12-04T08:54:16.5007717Z /home/runner/_work/_temp/f8be5534-de81-4bb1-a4ee-7faece589516.sh: line 5: aws: command not found 2025-12-04T08:54:16.5008875Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:16.5104255Z Error: Cannot perform an interactive login from a non TTY device 2025-12-04T08:54:16.5115735Z + sleep 2 2025-12-04T08:54:18.5127123Z + login 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:18.5129138Z + aws ecr get-login-password --region us-east-1 2025-12-04T08:54:18.5129390Z /home/runner/_work/_temp/f8be5534-de81-4bb1-a4ee-7faece589516.sh: line 5: aws: command not found 2025-12-04T08:54:18.5130342Z + docker login -u AWS --password-stdin 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T08:54:18.5189403Z Error: Cannot perform an interactive login from a non TTY device 2025-12-04T08:54:18.5202398Z ++ docker manifest inspect 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:18.5202740Z ++ jq '[.layers[].size, .config.size] | add / 1024 / 1024' 2025-12-04T08:54:19.8574317Z + IMAGE_SIZE=18171.470620155334 2025-12-04T08:54:19.8574569Z + echo 'Compressed size of image in MB: 18171.470620155334' 2025-12-04T08:54:19.8574744Z + set -e 2025-12-04T08:54:19.8575048Z + docker inspect --type=image 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:19.8575380Z Compressed size of image in MB: 18171.470620155334 2025-12-04T08:54:19.8665529Z + retry docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:19.8665980Z + docker pull 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T08:54:20.9299521Z pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a: Pulling from pytorch/ci-image 2025-12-04T08:54:20.9300298Z 63e5bc7682b8: Pulling fs layer 2025-12-04T08:54:20.9300674Z 835841cca3b7: Pulling fs layer 2025-12-04T08:54:20.9300981Z aac69780afc8: Pulling fs layer 2025-12-04T08:54:20.9301281Z 029495b23122: Pulling fs layer 2025-12-04T08:54:20.9301585Z d0fb85b00833: Pulling fs layer 2025-12-04T08:54:20.9301886Z 59b639308833: Pulling fs layer 2025-12-04T08:54:20.9302182Z dc112c89d57a: Pulling fs layer 2025-12-04T08:54:20.9302475Z 522eab2402e5: Pulling fs layer 2025-12-04T08:54:20.9302772Z 2b5a11b41761: Pulling fs layer 2025-12-04T08:54:20.9303064Z 9681563a88ff: Pulling fs layer 2025-12-04T08:54:20.9303355Z 73e33534e9eb: Pulling fs layer 2025-12-04T08:54:20.9303650Z 5bfdaeb5578d: Pulling fs layer 2025-12-04T08:54:20.9303876Z c07d27e4d3a5: Pulling fs layer 2025-12-04T08:54:20.9304088Z b21856d1bf42: Pulling fs layer 2025-12-04T08:54:20.9304318Z cb19d84867e4: Pulling fs layer 2025-12-04T08:54:20.9304530Z 8165374f8dcc: Pulling fs layer 2025-12-04T08:54:20.9304742Z 1aecc77354ce: Pulling fs layer 2025-12-04T08:54:20.9304954Z 465d3fd643aa: Pulling fs layer 2025-12-04T08:54:20.9305857Z 9681563a88ff: Waiting 2025-12-04T08:54:20.9306062Z 6c503e779d6f: Pulling fs layer 2025-12-04T08:54:20.9306305Z 73e33534e9eb: Waiting 2025-12-04T08:54:20.9306737Z 5bfdaeb5578d: Waiting 2025-12-04T08:54:20.9307221Z 59b639308833: Waiting 2025-12-04T08:54:20.9307706Z dc112c89d57a: Waiting 2025-12-04T08:54:20.9311163Z f7e9a021f0ee: Pulling fs layer 2025-12-04T08:54:20.9311478Z 8e023b349080: Pulling fs layer 2025-12-04T08:54:20.9311804Z 522eab2402e5: Waiting 2025-12-04T08:54:20.9312111Z 8188df80e595: Pulling fs layer 2025-12-04T08:54:20.9313691Z 2b5a11b41761: Waiting 2025-12-04T08:54:20.9314237Z 3c2c2f8c74bf: Pulling fs layer 2025-12-04T08:54:20.9314756Z 2aa7784fbe33: Pulling fs layer 2025-12-04T08:54:20.9315085Z 8165374f8dcc: Waiting 2025-12-04T08:54:20.9315796Z 2b3b5215d3eb: Pulling fs layer 2025-12-04T08:54:20.9316098Z 1aecc77354ce: Waiting 2025-12-04T08:54:20.9327980Z 99b1f1ea3e85: Pulling fs layer 2025-12-04T08:54:20.9328175Z 465d3fd643aa: Waiting 2025-12-04T08:54:20.9328316Z 6c503e779d6f: Waiting 2025-12-04T08:54:20.9328461Z 18d6daba0a57: Pulling fs layer 2025-12-04T08:54:20.9328599Z 8e023b349080: Waiting 2025-12-04T08:54:20.9328733Z 5277f2a503eb: Pulling fs layer 2025-12-04T08:54:20.9328882Z f7e9a021f0ee: Waiting 2025-12-04T08:54:20.9329018Z 3198a9717aac: Pulling fs layer 2025-12-04T08:54:20.9329157Z c07d27e4d3a5: Waiting 2025-12-04T08:54:20.9329296Z 99a4918e5808: Pulling fs layer 2025-12-04T08:54:20.9329445Z 15bb11dfc6ac: Pulling fs layer 2025-12-04T08:54:20.9329593Z bd87c8766e90: Pulling fs layer 2025-12-04T08:54:20.9329736Z cb19d84867e4: Waiting 2025-12-04T08:54:20.9329863Z 3c2c2f8c74bf: Waiting 2025-12-04T08:54:20.9329991Z b21856d1bf42: Waiting 2025-12-04T08:54:20.9330115Z 029495b23122: Waiting 2025-12-04T08:54:20.9330247Z 1969e15d0c13: Pulling fs layer 2025-12-04T08:54:20.9330407Z 24a03847d382: Pulling fs layer 2025-12-04T08:54:20.9330552Z 3198a9717aac: Waiting 2025-12-04T08:54:20.9330684Z 816e2e34e018: Pulling fs layer 2025-12-04T08:54:20.9330827Z 2aa7784fbe33: Waiting 2025-12-04T08:54:20.9330962Z b168858b8537: Pulling fs layer 2025-12-04T08:54:20.9331102Z 2b3b5215d3eb: Waiting 2025-12-04T08:54:20.9331230Z 99a4918e5808: Waiting 2025-12-04T08:54:20.9331354Z 18d6daba0a57: Waiting 2025-12-04T08:54:20.9331483Z bd87c8766e90: Waiting 2025-12-04T08:54:20.9331606Z 99b1f1ea3e85: Waiting 2025-12-04T08:54:20.9331743Z 6b8d5ff02e26: Pulling fs layer 2025-12-04T08:54:20.9331894Z 4e3b10a5dd6a: Pulling fs layer 2025-12-04T08:54:20.9332038Z 15bb11dfc6ac: Waiting 2025-12-04T08:54:20.9332168Z 6b8d5ff02e26: Waiting 2025-12-04T08:54:20.9332292Z b168858b8537: Waiting 2025-12-04T08:54:20.9332418Z 816e2e34e018: Waiting 2025-12-04T08:54:20.9332540Z 24a03847d382: Waiting 2025-12-04T08:54:20.9332675Z 3092fab73b59: Pulling fs layer 2025-12-04T08:54:20.9332818Z 1969e15d0c13: Waiting 2025-12-04T08:54:20.9332948Z 20020dd28a15: Pulling fs layer 2025-12-04T08:54:20.9333090Z 5277f2a503eb: Waiting 2025-12-04T08:54:20.9333214Z 3092fab73b59: Waiting 2025-12-04T08:54:20.9333345Z ae5280ce969d: Pulling fs layer 2025-12-04T08:54:20.9333491Z 8188df80e595: Waiting 2025-12-04T08:54:20.9333619Z 4f4fb700ef54: Pulling fs layer 2025-12-04T08:54:20.9333769Z fe17d9eb0fd2: Pulling fs layer 2025-12-04T08:54:20.9333916Z ae5280ce969d: Waiting 2025-12-04T08:54:20.9334045Z 20020dd28a15: Waiting 2025-12-04T08:54:20.9334176Z a51e0dab2d59: Pulling fs layer 2025-12-04T08:54:20.9334323Z fe17d9eb0fd2: Waiting 2025-12-04T08:54:20.9334460Z 6eb176cefd72: Pulling fs layer 2025-12-04T08:54:20.9334602Z a51e0dab2d59: Waiting 2025-12-04T08:54:20.9334734Z e7b8cf2e8d5a: Pulling fs layer 2025-12-04T08:54:20.9334867Z 4f4fb700ef54: Waiting 2025-12-04T08:54:20.9334985Z ef3a5060abce: Pulling fs layer 2025-12-04T08:54:20.9335112Z 6eb176cefd72: Waiting 2025-12-04T08:54:20.9335214Z e7b8cf2e8d5a: Waiting 2025-12-04T08:54:20.9335326Z a6f4ec14b42b: Pulling fs layer 2025-12-04T08:54:20.9335446Z 7e5a0c956cfb: Pulling fs layer 2025-12-04T08:54:20.9335560Z a6f4ec14b42b: Waiting 2025-12-04T08:54:20.9335668Z b4f78730cfe7: Pulling fs layer 2025-12-04T08:54:20.9335947Z ef3a5060abce: Waiting 2025-12-04T08:54:20.9336221Z 081028f24389: Pulling fs layer 2025-12-04T08:54:20.9336341Z a534dcf4b9a9: Pulling fs layer 2025-12-04T08:54:20.9336452Z 7e5a0c956cfb: Waiting 2025-12-04T08:54:20.9336560Z 2e77500302cc: Pulling fs layer 2025-12-04T08:54:20.9336686Z b4f78730cfe7: Waiting 2025-12-04T08:54:20.9336831Z bc08246bb4ba: Pulling fs layer 2025-12-04T08:54:20.9336947Z a534dcf4b9a9: Waiting 2025-12-04T08:54:20.9337049Z bc08246bb4ba: Waiting 2025-12-04T08:54:20.9337151Z 2e77500302cc: Waiting 2025-12-04T08:54:20.9337250Z 081028f24389: Waiting 2025-12-04T08:54:20.9337355Z ff0c473ca120: Pulling fs layer 2025-12-04T08:54:20.9337478Z 6bbc14b250ef: Pulling fs layer 2025-12-04T08:54:21.5239353Z 835841cca3b7: Download complete 2025-12-04T08:54:22.0974882Z 029495b23122: Download complete 2025-12-04T08:54:22.5877196Z 63e5bc7682b8: Download complete 2025-12-04T08:54:23.0949219Z 63e5bc7682b8: Pull complete 2025-12-04T08:54:23.1036483Z 835841cca3b7: Pull complete 2025-12-04T08:54:23.1761623Z 59b639308833: Download complete 2025-12-04T08:54:23.7821129Z 522eab2402e5: Verifying Checksum 2025-12-04T08:54:23.7821532Z 522eab2402e5: Download complete 2025-12-04T08:54:24.3666682Z 2b5a11b41761: Verifying Checksum 2025-12-04T08:54:24.3667247Z 2b5a11b41761: Download complete 2025-12-04T08:54:24.9483987Z 9681563a88ff: Verifying Checksum 2025-12-04T08:54:24.9484244Z 9681563a88ff: Download complete 2025-12-04T08:54:26.0147692Z dc112c89d57a: Verifying Checksum 2025-12-04T08:54:26.0148141Z dc112c89d57a: Download complete 2025-12-04T08:54:26.6261042Z 5bfdaeb5578d: Verifying Checksum 2025-12-04T08:54:26.6261369Z 5bfdaeb5578d: Download complete 2025-12-04T08:54:27.5954306Z c07d27e4d3a5: Download complete 2025-12-04T08:54:28.1892278Z b21856d1bf42: Verifying Checksum 2025-12-04T08:54:28.1892587Z b21856d1bf42: Download complete 2025-12-04T08:54:28.6705012Z aac69780afc8: Verifying Checksum 2025-12-04T08:54:28.6705415Z aac69780afc8: Download complete 2025-12-04T08:54:28.7833618Z cb19d84867e4: Verifying Checksum 2025-12-04T08:54:28.7834010Z cb19d84867e4: Download complete 2025-12-04T08:54:29.2592797Z 8165374f8dcc: Verifying Checksum 2025-12-04T08:54:29.2592940Z 8165374f8dcc: Download complete 2025-12-04T08:54:29.8669775Z 465d3fd643aa: Verifying Checksum 2025-12-04T08:54:29.8670019Z 465d3fd643aa: Download complete 2025-12-04T08:54:30.4581393Z 6c503e779d6f: Download complete 2025-12-04T08:54:31.0560567Z f7e9a021f0ee: Verifying Checksum 2025-12-04T08:54:31.0560778Z f7e9a021f0ee: Download complete 2025-12-04T08:54:31.6357825Z 8e023b349080: Download complete 2025-12-04T08:54:32.7296736Z aac69780afc8: Pull complete 2025-12-04T08:54:32.7349190Z 029495b23122: Pull complete 2025-12-04T08:54:32.7408681Z d0fb85b00833: Pull complete 2025-12-04T08:54:32.7446824Z 59b639308833: Pull complete 2025-12-04T08:54:33.7944468Z dc112c89d57a: Pull complete 2025-12-04T08:54:33.7985409Z 522eab2402e5: Pull complete 2025-12-04T08:54:33.8027974Z 2b5a11b41761: Pull complete 2025-12-04T08:54:33.8064034Z 9681563a88ff: Pull complete 2025-12-04T08:54:44.7948059Z 1aecc77354ce: Verifying Checksum 2025-12-04T08:54:44.7948278Z 1aecc77354ce: Download complete 2025-12-04T08:54:45.4296939Z 3c2c2f8c74bf: Download complete 2025-12-04T08:54:46.0498968Z 2aa7784fbe33: Download complete 2025-12-04T08:54:46.6587755Z 2b3b5215d3eb: Verifying Checksum 2025-12-04T08:54:46.6588084Z 2b3b5215d3eb: Download complete 2025-12-04T08:55:08.2344140Z 99b1f1ea3e85: Verifying Checksum 2025-12-04T08:55:08.2345005Z 99b1f1ea3e85: Download complete 2025-12-04T08:55:08.8648531Z 18d6daba0a57: Verifying Checksum 2025-12-04T08:55:08.8678078Z 18d6daba0a57: Download complete 2025-12-04T08:55:09.4574777Z 5277f2a503eb: Verifying Checksum 2025-12-04T08:55:09.4575012Z 5277f2a503eb: Download complete 2025-12-04T08:55:10.0407568Z 3198a9717aac: Download complete 2025-12-04T08:55:10.6598635Z 99a4918e5808: Verifying Checksum 2025-12-04T08:55:10.6599035Z 99a4918e5808: Download complete 2025-12-04T08:55:11.7711546Z 15bb11dfc6ac: Verifying Checksum 2025-12-04T08:55:11.7711797Z 15bb11dfc6ac: Download complete 2025-12-04T08:55:12.3410652Z bd87c8766e90: Verifying Checksum 2025-12-04T08:55:12.3411119Z bd87c8766e90: Download complete 2025-12-04T08:55:12.9259184Z 1969e15d0c13: Download complete 2025-12-04T08:55:14.6331912Z 24a03847d382: Verifying Checksum 2025-12-04T08:55:14.6332246Z 24a03847d382: Download complete 2025-12-04T08:55:15.2126927Z 816e2e34e018: Verifying Checksum 2025-12-04T08:55:15.2127157Z 816e2e34e018: Download complete 2025-12-04T08:55:15.7804762Z b168858b8537: Verifying Checksum 2025-12-04T08:55:15.7806027Z b168858b8537: Download complete 2025-12-04T08:55:17.2394541Z 6b8d5ff02e26: Verifying Checksum 2025-12-04T08:55:17.2395007Z 6b8d5ff02e26: Download complete 2025-12-04T08:55:17.8232809Z 4e3b10a5dd6a: Download complete 2025-12-04T08:55:18.3915512Z 3092fab73b59: Verifying Checksum 2025-12-04T08:55:18.3916579Z 3092fab73b59: Download complete 2025-12-04T08:55:18.9839349Z 20020dd28a15: Verifying Checksum 2025-12-04T08:55:18.9839738Z 20020dd28a15: Download complete 2025-12-04T08:55:19.5624805Z ae5280ce969d: Download complete 2025-12-04T08:55:19.8597074Z 4f4fb700ef54: Download complete 2025-12-04T08:55:20.4669796Z fe17d9eb0fd2: Verifying Checksum 2025-12-04T08:55:20.4670058Z fe17d9eb0fd2: Download complete 2025-12-04T08:55:21.0692073Z a51e0dab2d59: Download complete 2025-12-04T08:55:21.8109503Z 6eb176cefd72: Download complete 2025-12-04T08:55:22.3978504Z e7b8cf2e8d5a: Download complete 2025-12-04T08:55:22.9852700Z ef3a5060abce: Verifying Checksum 2025-12-04T08:55:22.9853188Z ef3a5060abce: Download complete 2025-12-04T08:55:24.1450123Z 7e5a0c956cfb: Verifying Checksum 2025-12-04T08:55:24.1450322Z 7e5a0c956cfb: Download complete 2025-12-04T09:03:11.2192650Z 73e33534e9eb: Verifying Checksum 2025-12-04T09:03:11.2193035Z 73e33534e9eb: Download complete 2025-12-04T09:03:11.8092773Z 081028f24389: Download complete 2025-12-04T09:03:12.3869352Z a534dcf4b9a9: Verifying Checksum 2025-12-04T09:03:12.3869512Z a534dcf4b9a9: Download complete 2025-12-04T09:03:17.6624405Z 2e77500302cc: Verifying Checksum 2025-12-04T09:03:17.6624818Z 2e77500302cc: Download complete 2025-12-04T09:03:18.3081019Z bc08246bb4ba: Verifying Checksum 2025-12-04T09:03:18.3081237Z bc08246bb4ba: Download complete 2025-12-04T09:03:18.8857846Z ff0c473ca120: Verifying Checksum 2025-12-04T09:03:18.8858218Z ff0c473ca120: Download complete 2025-12-04T09:03:21.0701254Z 6bbc14b250ef: Verifying Checksum 2025-12-04T09:03:21.0701481Z 6bbc14b250ef: Download complete 2025-12-04T09:03:34.0404798Z 73e33534e9eb: Pull complete 2025-12-04T09:03:34.0460706Z 5bfdaeb5578d: Pull complete 2025-12-04T09:03:34.0631145Z c07d27e4d3a5: Pull complete 2025-12-04T09:03:34.0677543Z b21856d1bf42: Pull complete 2025-12-04T09:03:34.0719363Z cb19d84867e4: Pull complete 2025-12-04T09:03:34.0767685Z 8165374f8dcc: Pull complete 2025-12-04T09:03:37.8377998Z 1aecc77354ce: Pull complete 2025-12-04T09:03:37.8436395Z 465d3fd643aa: Pull complete 2025-12-04T09:03:37.8565259Z 6c503e779d6f: Pull complete 2025-12-04T09:03:37.8681370Z f7e9a021f0ee: Pull complete 2025-12-04T09:03:37.8731496Z 8e023b349080: Pull complete 2025-12-04T09:17:33.7054987Z b4f78730cfe7: Verifying Checksum 2025-12-04T09:17:33.7055450Z b4f78730cfe7: Download complete 2025-12-04T09:27:11.8406473Z 8188df80e595: Verifying Checksum 2025-12-04T09:27:11.8406860Z 8188df80e595: Download complete 2025-12-04T09:27:58.1505471Z 8188df80e595: Pull complete 2025-12-04T09:27:58.1634981Z 3c2c2f8c74bf: Pull complete 2025-12-04T09:27:58.1695259Z 2aa7784fbe33: Pull complete 2025-12-04T09:27:58.1743236Z 2b3b5215d3eb: Pull complete 2025-12-04T09:28:03.5541702Z 99b1f1ea3e85: Pull complete 2025-12-04T09:28:03.5633940Z 18d6daba0a57: Pull complete 2025-12-04T09:28:03.5690772Z 5277f2a503eb: Pull complete 2025-12-04T09:28:03.5733970Z 3198a9717aac: Pull complete 2025-12-04T09:28:03.5778328Z 99a4918e5808: Pull complete 2025-12-04T09:28:03.6099297Z 15bb11dfc6ac: Pull complete 2025-12-04T09:28:03.6150267Z bd87c8766e90: Pull complete 2025-12-04T09:28:03.6197318Z 1969e15d0c13: Pull complete 2025-12-04T09:28:03.8343373Z 24a03847d382: Pull complete 2025-12-04T09:28:03.8390640Z 816e2e34e018: Pull complete 2025-12-04T09:28:03.8434678Z b168858b8537: Pull complete 2025-12-04T09:28:03.9426490Z 6b8d5ff02e26: Pull complete 2025-12-04T09:28:03.9514205Z 4e3b10a5dd6a: Pull complete 2025-12-04T09:28:03.9614303Z 3092fab73b59: Pull complete 2025-12-04T09:28:03.9644154Z 20020dd28a15: Pull complete 2025-12-04T09:28:03.9692998Z ae5280ce969d: Pull complete 2025-12-04T09:28:03.9737181Z 4f4fb700ef54: Pull complete 2025-12-04T09:28:03.9794750Z fe17d9eb0fd2: Pull complete 2025-12-04T09:28:03.9846623Z a51e0dab2d59: Pull complete 2025-12-04T09:28:03.9973964Z 6eb176cefd72: Pull complete 2025-12-04T09:28:04.0022306Z e7b8cf2e8d5a: Pull complete 2025-12-04T09:28:04.0059370Z ef3a5060abce: Pull complete 2025-12-04T09:28:04.0152852Z a6f4ec14b42b: Pull complete 2025-12-04T09:28:04.0191263Z 7e5a0c956cfb: Pull complete 2025-12-04T09:28:42.8532272Z b4f78730cfe7: Pull complete 2025-12-04T09:28:42.8584768Z 081028f24389: Pull complete 2025-12-04T09:28:42.8629967Z a534dcf4b9a9: Pull complete 2025-12-04T09:28:45.7384837Z 2e77500302cc: Pull complete 2025-12-04T09:28:45.7437679Z bc08246bb4ba: Pull complete 2025-12-04T09:28:45.7523008Z ff0c473ca120: Pull complete 2025-12-04T09:28:46.3805745Z 6bbc14b250ef: Pull complete 2025-12-04T09:28:46.3823227Z Digest: sha256:5e190224966743059cf8506170eaec525eada34e38cf646e02d1dbeadfe5a366 2025-12-04T09:28:46.3833183Z Status: Downloaded newer image for 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:28:46.3839648Z 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:28:46.3892631Z Prepare all required actions 2025-12-04T09:28:46.3920821Z ##[group]Run ./.github/actions/get-workflow-job-id 2025-12-04T09:28:46.3921049Z with: 2025-12-04T09:28:46.3921423Z github-token: *** 2025-12-04T09:28:46.3921559Z env: 2025-12-04T09:28:46.3921684Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:28:46.3921869Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:28:46.3922095Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:28:46.3922322Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:28:46.3922810Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:28:46.3923299Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:28:46.3923449Z AWS_REGION: us-east-1 2025-12-04T09:28:46.3923610Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:28:46.3923841Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:28:46.3926592Z AWS_SESSION_TOKEN: *** 2025-12-04T09:28:46.3926734Z ##[endgroup] 2025-12-04T09:28:46.3935753Z ##[group]Run set -eux 2025-12-04T09:28:46.3935876Z set -eux 2025-12-04T09:28:46.3936048Z python3 .github/scripts/get_workflow_job_id.py "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-12-04T09:28:46.3940684Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:28:46.3940831Z env: 2025-12-04T09:28:46.3940925Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:28:46.3941059Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:28:46.3941233Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:28:46.3941400Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:28:46.3941785Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:28:46.3942152Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:28:46.3942267Z AWS_REGION: us-east-1 2025-12-04T09:28:46.3942445Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:28:46.3942598Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:28:46.3944790Z AWS_SESSION_TOKEN: *** 2025-12-04T09:28:46.3944948Z GITHUB_TOKEN: *** 2025-12-04T09:28:46.3945043Z ##[endgroup] 2025-12-04T09:28:46.3970967Z + python3 .github/scripts/get_workflow_job_id.py 19922849170 linux.rocm.gpu.gfx942.1.b-gwk9b-runner-68hm6 2025-12-04T09:28:47.4433872Z Setting output job-id=57116213149 2025-12-04T09:28:47.4434689Z Setting output job-name=linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable) 2025-12-04T09:28:47.4561078Z Prepare all required actions 2025-12-04T09:28:47.4561313Z Getting action download info 2025-12-04T09:28:47.8516152Z Download action repository 'seemethere/download-artifact-s3@v4' (SHA:1da556a7aa0a088e3153970611f6c432d58e80e6) 2025-12-04T09:28:48.9802449Z Download action repository 'actions/download-artifact@v4' (SHA:d3f86a106a0bac45b974a628896c90dbdf5c8093) 2025-12-04T09:28:50.1016646Z ##[group]Run ./.github/actions/download-build-artifacts 2025-12-04T09:28:50.1016872Z with: 2025-12-04T09:28:50.1016980Z name: linux-jammy-rocm-py3.10 2025-12-04T09:28:50.1017102Z s3-bucket: gha-artifacts 2025-12-04T09:28:50.1017215Z env: 2025-12-04T09:28:50.1017310Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:28:50.1017444Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:28:50.1017621Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:28:50.1017790Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:28:50.1018195Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:28:50.1018574Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:28:50.1018691Z AWS_REGION: us-east-1 2025-12-04T09:28:50.1018853Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:28:50.1019014Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:28:50.1021200Z AWS_SESSION_TOKEN: *** 2025-12-04T09:28:50.1021306Z ##[endgroup] 2025-12-04T09:28:50.1053594Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:28:50.1053730Z with: 2025-12-04T09:28:50.1053833Z name: linux-jammy-rocm-py3.10 2025-12-04T09:28:50.1074549Z s3-bucket: gha-artifacts 2025-12-04T09:28:50.1074663Z region: us-east-1 2025-12-04T09:28:50.1074764Z env: 2025-12-04T09:28:50.1074853Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:28:50.1074986Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:28:50.1075164Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:28:50.1075335Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:28:50.1075722Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:28:50.1076098Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:28:50.1076211Z AWS_REGION: us-east-1 2025-12-04T09:28:50.1076349Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:28:50.1076498Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:28:50.1078737Z AWS_SESSION_TOKEN: *** 2025-12-04T09:28:50.1078840Z ##[endgroup] 2025-12-04T09:28:50.3837158Z (node:17243) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:28:50.3837393Z 2025-12-04T09:28:50.3837477Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:28:50.3837789Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:28:50.3838023Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:28:50.6572713Z Found 1 objects with prefix pytorch/pytorch/19922849170/linux-jammy-rocm-py3.10/ 2025-12-04T09:28:50.6573013Z Starting download (1/1): /home/runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:31:18.0057956Z Finished download (1/1): /home/runner/_work/pytorch/pytorch/artifacts.zip 2025-12-04T09:31:18.0062236Z Artifact download has finished successfully 2025-12-04T09:31:18.0241943Z ##[group]Run unzip -o artifacts.zip 2025-12-04T09:31:18.0242127Z unzip -o artifacts.zip 2025-12-04T09:31:18.0246311Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:31:18.0246469Z env: 2025-12-04T09:31:18.0246570Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:18.0246712Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:18.0247159Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:18.0247336Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:18.0247739Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:18.0248129Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:18.0248448Z AWS_REGION: us-east-1 2025-12-04T09:31:18.0248641Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:18.0248802Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:18.0251078Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:18.0251190Z ##[endgroup] 2025-12-04T09:31:18.0286359Z Archive: artifacts.zip 2025-12-04T09:31:18.0287105Z creating: dist/ 2025-12-04T09:31:18.0381635Z inflating: dist/.ninja_log 2025-12-04T09:31:21.1798603Z inflating: dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl 2025-12-04T09:31:21.1801730Z creating: build/ 2025-12-04T09:31:21.1801921Z creating: build/custom_test_artifacts/ 2025-12-04T09:31:21.1802098Z creating: build/custom_test_artifacts/custom-op-build/ 2025-12-04T09:31:21.1802290Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/ 2025-12-04T09:31:21.1802532Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:31:21.1802790Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:31:21.1803045Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/ 2025-12-04T09:31:21.1803291Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:31:21.1803547Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:31:21.1803798Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:31:21.1804091Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:31:21.1804383Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:31:21.1804658Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:31:21.1804923Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:31:21.1805183Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:31:21.1805674Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:31:21.1806465Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:31:21.1806968Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:31:21.1810390Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:31:21.1811167Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:31:21.1811452Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:31:21.1811679Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:31:21.1811919Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/cmake.check_cache 2025-12-04T09:31:21.1812165Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/ 2025-12-04T09:31:21.1812432Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.ts 2025-12-04T09:31:21.1812734Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/compiler_depend.make 2025-12-04T09:31:21.1813442Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/depend.make 2025-12-04T09:31:21.1813712Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/link.txt 2025-12-04T09:31:21.1813990Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/cmake_clean.cmake 2025-12-04T09:31:21.1814270Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/build.make 2025-12-04T09:31:21.1814719Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/DependInfo.cmake 2025-12-04T09:31:21.1814999Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/flags.make 2025-12-04T09:31:21.1815278Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/progress.make 2025-12-04T09:31:21.1825355Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o.d 2025-12-04T09:31:21.1941464Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/custom_ops.dir/op.cpp.o 2025-12-04T09:31:21.1945148Z creating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/ 2025-12-04T09:31:21.1945457Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.ts 2025-12-04T09:31:21.1945787Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/compiler_depend.make 2025-12-04T09:31:21.1946105Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/depend.make 2025-12-04T09:31:21.1946398Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/link.txt 2025-12-04T09:31:21.1946699Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/cmake_clean.cmake 2025-12-04T09:31:21.1947033Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/build.make 2025-12-04T09:31:21.1947336Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/DependInfo.cmake 2025-12-04T09:31:21.1947636Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/flags.make 2025-12-04T09:31:21.1947930Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/progress.make 2025-12-04T09:31:21.1960570Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o.d 2025-12-04T09:31:21.2006983Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/test_custom_ops.dir/test_custom_ops.cpp.o 2025-12-04T09:31:21.2007321Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:31:21.2007618Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:31:21.2007887Z extracting: build/custom_test_artifacts/custom-op-build/CMakeFiles/progress.marks 2025-12-04T09:31:21.2008138Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile2 2025-12-04T09:31:21.2008552Z inflating: build/custom_test_artifacts/custom-op-build/CMakeFiles/Makefile.cmake 2025-12-04T09:31:21.2008800Z inflating: build/custom_test_artifacts/custom-op-build/hipblaslt_test_outer_vec.cc 2025-12-04T09:31:21.2009049Z inflating: build/custom_test_artifacts/custom-op-build/hipblaslt_test_vec_ext.cc 2025-12-04T09:31:21.2009901Z inflating: build/custom_test_artifacts/custom-op-build/CMakeCache.txt 2025-12-04T09:31:21.2010242Z inflating: build/custom_test_artifacts/custom-op-build/Makefile 2025-12-04T09:31:21.2010639Z inflating: build/custom_test_artifacts/custom-op-build/cmake_install.cmake 2025-12-04T09:31:21.2103124Z inflating: build/custom_test_artifacts/custom-op-build/libcustom_ops.so 2025-12-04T09:31:21.2133063Z inflating: build/custom_test_artifacts/custom-op-build/test_custom_ops 2025-12-04T09:31:21.2133361Z creating: build/custom_test_artifacts/jit-hook-build/ 2025-12-04T09:31:21.2134201Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/ 2025-12-04T09:31:21.2134505Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:31:21.2135621Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:31:21.2135972Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/ 2025-12-04T09:31:21.2136386Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:31:21.2136776Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:31:21.2137116Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:31:21.2137545Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:31:21.2138319Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:31:21.2138699Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:31:21.2139058Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:31:21.2139411Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:31:21.2140224Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:31:21.2140902Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:31:21.2141317Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:31:21.2142355Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:31:21.2143078Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:31:21.2143417Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:31:21.2143682Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:31:21.2143968Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/cmake.check_cache 2025-12-04T09:31:21.2144262Z creating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/ 2025-12-04T09:31:21.2144594Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.ts 2025-12-04T09:31:21.2144969Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/compiler_depend.make 2025-12-04T09:31:21.2145329Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/depend.make 2025-12-04T09:31:21.2145664Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/link.txt 2025-12-04T09:31:21.2146019Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/cmake_clean.cmake 2025-12-04T09:31:21.2146369Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/build.make 2025-12-04T09:31:21.2146716Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/DependInfo.cmake 2025-12-04T09:31:21.2147119Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/flags.make 2025-12-04T09:31:21.2147462Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/progress.make 2025-12-04T09:31:21.2157839Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o.d 2025-12-04T09:31:21.2192370Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/test_jit_hooks.dir/test_jit_hooks.cpp.o 2025-12-04T09:31:21.2192704Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:31:21.2193028Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:31:21.2193292Z extracting: build/custom_test_artifacts/jit-hook-build/CMakeFiles/progress.marks 2025-12-04T09:31:21.2193532Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile2 2025-12-04T09:31:21.2194211Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeFiles/Makefile.cmake 2025-12-04T09:31:21.2194490Z inflating: build/custom_test_artifacts/jit-hook-build/hipblaslt_test_outer_vec.cc 2025-12-04T09:31:21.2195095Z inflating: build/custom_test_artifacts/jit-hook-build/hipblaslt_test_vec_ext.cc 2025-12-04T09:31:21.2195457Z inflating: build/custom_test_artifacts/jit-hook-build/CMakeCache.txt 2025-12-04T09:31:21.2195835Z inflating: build/custom_test_artifacts/jit-hook-build/Makefile 2025-12-04T09:31:21.2196111Z inflating: build/custom_test_artifacts/jit-hook-build/cmake_install.cmake 2025-12-04T09:31:21.2225876Z inflating: build/custom_test_artifacts/jit-hook-build/test_jit_hooks 2025-12-04T09:31:21.2226083Z creating: build/custom_test_artifacts/custom-backend-build/ 2025-12-04T09:31:21.2226293Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/ 2025-12-04T09:31:21.2226530Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/pkgRedirects/ 2025-12-04T09:31:21.2228588Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeConfigureLog.yaml 2025-12-04T09:31:21.2228858Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/ 2025-12-04T09:31:21.2229117Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeSystem.cmake 2025-12-04T09:31:21.2229394Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/ 2025-12-04T09:31:21.2229666Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/tmp/ 2025-12-04T09:31:21.2230472Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/CMakeCCompilerId.c 2025-12-04T09:31:21.2231241Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdC/a.out 2025-12-04T09:31:21.2231533Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCCompiler.cmake 2025-12-04T09:31:21.2231819Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/ 2025-12-04T09:31:21.2232096Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/tmp/ 2025-12-04T09:31:21.2233115Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/CMakeCXXCompilerId.cpp 2025-12-04T09:31:21.2233877Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CompilerIdCXX/a.out 2025-12-04T09:31:21.2234265Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeCXXCompiler.cmake 2025-12-04T09:31:21.2238364Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_C.bin 2025-12-04T09:31:21.2239112Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/3.31.6/CMakeDetermineCompilerABI_CXX.bin 2025-12-04T09:31:21.2239412Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeScratch/ 2025-12-04T09:31:21.2239655Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeTmp/ 2025-12-04T09:31:21.2239909Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/cmake.check_cache 2025-12-04T09:31:21.2240172Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/ 2025-12-04T09:31:21.2240467Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.ts 2025-12-04T09:31:21.2240928Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/compiler_depend.make 2025-12-04T09:31:21.2241251Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/depend.make 2025-12-04T09:31:21.2241550Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/link.txt 2025-12-04T09:31:21.2241858Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/cmake_clean.cmake 2025-12-04T09:31:21.2242203Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/build.make 2025-12-04T09:31:21.2242514Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/DependInfo.cmake 2025-12-04T09:31:21.2242825Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/flags.make 2025-12-04T09:31:21.2243130Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/progress.make 2025-12-04T09:31:21.2244389Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o.d 2025-12-04T09:31:21.2320646Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/custom_backend.dir/custom_backend.cpp.o 2025-12-04T09:31:21.2323622Z creating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/ 2025-12-04T09:31:21.2323937Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.ts 2025-12-04T09:31:21.2324291Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/compiler_depend.make 2025-12-04T09:31:21.2324624Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/depend.make 2025-12-04T09:31:21.2324940Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/link.txt 2025-12-04T09:31:21.2325268Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/cmake_clean.cmake 2025-12-04T09:31:21.2325597Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/build.make 2025-12-04T09:31:21.2325923Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/DependInfo.cmake 2025-12-04T09:31:21.2326250Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/flags.make 2025-12-04T09:31:21.2326573Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/progress.make 2025-12-04T09:31:21.2338300Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o.d 2025-12-04T09:31:21.2368128Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/test_custom_backend.dir/test_custom_backend.cpp.o 2025-12-04T09:31:21.2368499Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/CMakeDirectoryInformation.cmake 2025-12-04T09:31:21.2368804Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/TargetDirectories.txt 2025-12-04T09:31:21.2369080Z extracting: build/custom_test_artifacts/custom-backend-build/CMakeFiles/progress.marks 2025-12-04T09:31:21.2369331Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile2 2025-12-04T09:31:21.2369717Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeFiles/Makefile.cmake 2025-12-04T09:31:21.2369984Z inflating: build/custom_test_artifacts/custom-backend-build/hipblaslt_test_outer_vec.cc 2025-12-04T09:31:21.2370242Z inflating: build/custom_test_artifacts/custom-backend-build/hipblaslt_test_vec_ext.cc 2025-12-04T09:31:21.2371050Z inflating: build/custom_test_artifacts/custom-backend-build/CMakeCache.txt 2025-12-04T09:31:21.2371439Z inflating: build/custom_test_artifacts/custom-backend-build/Makefile 2025-12-04T09:31:21.2371856Z inflating: build/custom_test_artifacts/custom-backend-build/cmake_install.cmake 2025-12-04T09:31:21.2426305Z inflating: build/custom_test_artifacts/custom-backend-build/libcustom_backend.so 2025-12-04T09:31:21.2447304Z inflating: build/custom_test_artifacts/custom-backend-build/test_custom_backend 2025-12-04T09:31:21.2447490Z creating: build/lib/ 2025-12-04T09:31:21.2493237Z inflating: build/lib/libprotobuf-lite.a 2025-12-04T09:31:21.2754727Z inflating: build/lib/libprotobuf.a 2025-12-04T09:31:21.3029321Z inflating: build/lib/libprotoc.a 2025-12-04T09:31:21.3034652Z inflating: build/lib/libpthreadpool.a 2025-12-04T09:31:21.3038926Z inflating: build/lib/libcpuinfo.a 2025-12-04T09:31:21.3042991Z inflating: build/lib/libcpuinfo_internals.a 2025-12-04T09:31:21.3044276Z inflating: build/lib/libclog.a 2025-12-04T09:31:21.3054669Z inflating: build/lib/libpytorch_qnnpack.a 2025-12-04T09:31:21.3055527Z inflating: build/lib/libnnpack_reference_layers.a 2025-12-04T09:31:21.3065344Z inflating: build/lib/libnnpack.a 2025-12-04T09:31:21.3186257Z inflating: build/lib/libmicrokernels-prod.a 2025-12-04T09:31:21.3673258Z inflating: build/lib/libmicrokernels-all.a 2025-12-04T09:31:21.3711530Z inflating: build/lib/libgtest.a 2025-12-04T09:31:21.3720863Z inflating: build/lib/libgmock.a 2025-12-04T09:31:21.3721076Z inflating: build/lib/libgtest_main.a 2025-12-04T09:31:21.3721277Z inflating: build/lib/libgmock_main.a 2025-12-04T09:31:21.3770514Z inflating: build/lib/libXNNPACK.a 2025-12-04T09:31:21.3811855Z inflating: build/lib/libbenchmark.a 2025-12-04T09:31:21.3812079Z inflating: build/lib/libbenchmark_main.a 2025-12-04T09:31:21.3812286Z inflating: build/lib/libjitprofiling.a 2025-12-04T09:31:21.3816622Z inflating: build/lib/libittnotify.a 2025-12-04T09:31:21.3852867Z inflating: build/lib/libasmjit.a 2025-12-04T09:31:21.4524369Z inflating: build/lib/libfbgemm.a 2025-12-04T09:31:21.4540406Z inflating: build/lib/libtensorpipe_uv.a 2025-12-04T09:31:21.4852484Z inflating: build/lib/libtensorpipe.a 2025-12-04T09:31:21.4927299Z inflating: build/lib/libgloo.a 2025-12-04T09:31:21.4953113Z inflating: build/lib/libonnx_proto.a 2025-12-04T09:31:21.5189799Z inflating: build/lib/libgloo_hip.a 2025-12-04T09:31:21.5609705Z inflating: build/lib/libonnx.a 2025-12-04T09:31:22.1442180Z inflating: build/lib/libdnnl.a 2025-12-04T09:31:22.1454529Z inflating: build/lib/libfmt.a 2025-12-04T09:31:22.1639920Z inflating: build/lib/libkineto.a 2025-12-04T09:31:22.1704668Z inflating: build/lib/libc10.so 2025-12-04T09:31:22.1705381Z inflating: build/lib/libtorch_global_deps.so 2025-12-04T09:31:22.1706155Z inflating: build/lib/libcaffe2_nvrtc.so 2025-12-04T09:31:22.1731413Z inflating: build/lib/libc10_hip.so 2025-12-04T09:31:22.2020498Z inflating: build/lib/libfbgemm_genai.a 2025-12-04T09:31:23.9919132Z inflating: build/lib/libtorch_cpu.so 2025-12-04T09:31:23.9921225Z inflating: build/lib/libshm.so 2025-12-04T09:31:24.8620097Z inflating: build/lib/libtorch_hip.so 2025-12-04T09:31:24.8620593Z inflating: build/lib/libtorch.so 2025-12-04T09:31:24.8631364Z inflating: build/lib/libjitbackend_test.so 2025-12-04T09:31:24.8644625Z inflating: build/lib/libbackend_with_compiler.so 2025-12-04T09:31:24.8684210Z inflating: build/lib/libtorchbind_test.so 2025-12-04T09:31:24.8698827Z inflating: build/lib/libaoti_custom_ops.so 2025-12-04T09:31:25.0042471Z inflating: build/lib/libtorch_python.so 2025-12-04T09:31:25.0062136Z inflating: build/lib/libnnapi_backend.so 2025-12-04T09:31:25.0062428Z creating: build/bin/ 2025-12-04T09:31:25.0062659Z creating: build/bin/CMakeFiles/ 2025-12-04T09:31:25.0062927Z inflating: build/bin/cmake_install.cmake 2025-12-04T09:31:25.0063202Z inflating: build/bin/CTestTestfile.cmake 2025-12-04T09:31:25.0324365Z inflating: build/bin/protoc-3.13.0.0 2025-12-04T09:31:25.0598852Z inflating: build/bin/protoc 2025-12-04T09:31:25.0632858Z inflating: build/bin/c10_AllocatorConfig_test 2025-12-04T09:31:25.0663452Z inflating: build/bin/c10_CompileTimeFunctionPointer_test 2025-12-04T09:31:25.0694867Z inflating: build/bin/c10_DeviceGuard_test 2025-12-04T09:31:25.0726390Z inflating: build/bin/c10_Device_test 2025-12-04T09:31:25.0762337Z inflating: build/bin/c10_DispatchKeySet_test 2025-12-04T09:31:25.0795076Z inflating: build/bin/c10_Scalar_test 2025-12-04T09:31:25.0825161Z inflating: build/bin/c10_StreamGuard_test 2025-12-04T09:31:25.0859487Z inflating: build/bin/c10_SymInt_test 2025-12-04T09:31:25.0902366Z inflating: build/bin/c10_SizesAndStrides_test 2025-12-04T09:31:25.0936248Z inflating: build/bin/c10_Bitset_test 2025-12-04T09:31:25.0978811Z inflating: build/bin/c10_cow_test 2025-12-04T09:31:25.1014176Z inflating: build/bin/c10_InlineDeviceGuard_test 2025-12-04T09:31:25.1048871Z inflating: build/bin/c10_InlineStreamGuard_test 2025-12-04T09:31:25.1079078Z inflating: build/bin/c10_ArrayRef_test 2025-12-04T09:31:25.1109089Z inflating: build/bin/c10_ConstexprCrc_test 2025-12-04T09:31:25.1139362Z inflating: build/bin/c10_DeadlockDetection_test 2025-12-04T09:31:25.1171312Z inflating: build/bin/c10_IntrusiveList_test 2025-12-04T09:31:25.1202234Z inflating: build/bin/c10_Half_test 2025-12-04T09:31:25.1236792Z inflating: build/bin/c10_Enumerate_test 2025-12-04T09:31:25.1276801Z inflating: build/bin/c10_LeftRight_test 2025-12-04T09:31:25.1320697Z inflating: build/bin/c10_NetworkFlow_test 2025-12-04T09:31:25.1350725Z inflating: build/bin/c10_Semaphore_test 2025-12-04T09:31:25.1383930Z inflating: build/bin/c10_Synchronized_test 2025-12-04T09:31:25.1416961Z inflating: build/bin/c10_TypeIndex_test 2025-12-04T09:31:25.1450842Z inflating: build/bin/c10_ThreadLocal_test 2025-12-04T09:31:25.1482341Z inflating: build/bin/c10_accumulate_test 2025-12-04T09:31:25.1516136Z inflating: build/bin/c10_bfloat16_test 2025-12-04T09:31:25.1546291Z inflating: build/bin/c10_error_test 2025-12-04T09:31:25.1577131Z inflating: build/bin/c10_bit_cast_test 2025-12-04T09:31:25.1610471Z inflating: build/bin/c10_complex_test 2025-12-04T09:31:25.1642277Z inflating: build/bin/c10_exception_test 2025-12-04T09:31:25.1688566Z inflating: build/bin/c10_complex_math_test 2025-12-04T09:31:25.1722219Z inflating: build/bin/c10_flags_test 2025-12-04T09:31:25.1758358Z inflating: build/bin/c10_irange_test 2025-12-04T09:31:25.1789131Z inflating: build/bin/c10_generic_math_test 2025-12-04T09:31:25.1880879Z inflating: build/bin/c10_intrusive_ptr_test 2025-12-04T09:31:25.1916597Z inflating: build/bin/c10_logging_test 2025-12-04T09:31:25.1947041Z inflating: build/bin/c10_nofatal_test 2025-12-04T09:31:25.1979537Z inflating: build/bin/c10_lazy_test 2025-12-04T09:31:25.2016871Z inflating: build/bin/c10_ordered_preserving_dict_test 2025-12-04T09:31:25.2049159Z inflating: build/bin/c10_registry_test 2025-12-04T09:31:25.2089786Z inflating: build/bin/c10_ssize_test 2025-12-04T09:31:25.2139669Z inflating: build/bin/c10_optional_test 2025-12-04T09:31:25.2234921Z inflating: build/bin/c10_small_vector_test 2025-12-04T09:31:25.2269834Z inflating: build/bin/c10_string_util_test 2025-12-04T09:31:25.2300305Z inflating: build/bin/c10_tempfile_test 2025-12-04T09:31:25.2330353Z inflating: build/bin/c10_string_view_test 2025-12-04T09:31:25.2357068Z inflating: build/bin/c10_intrusive_ptr_benchmark 2025-12-04T09:31:25.2390826Z inflating: build/bin/c10_typeid_test 2025-12-04T09:31:25.2420793Z inflating: build/bin/c10_hip_HIPAssertionsTest_1_var_test 2025-12-04T09:31:25.2453397Z inflating: build/bin/c10_hip_HIPAssertionsTest_catches_stream 2025-12-04T09:31:25.2487633Z inflating: build/bin/c10_hip_HIPAssertionsTest_catches_thread_and_block_and_device 2025-12-04T09:31:25.2533882Z inflating: build/bin/c10_hip_HIPAssertionsTest_from_2_processes 2025-12-04T09:31:25.2567867Z inflating: build/bin/c10_hip_HIPAssertionsTest_multiple_writes_from_blocks_and_threads 2025-12-04T09:31:25.2596950Z inflating: build/bin/c10_hip_HIPAssertionsTest_multiple_writes_from_multiple_blocks 2025-12-04T09:31:25.2627029Z inflating: build/bin/c10_hip_HIPAssertionsTest_multiple_writes_from_same_block 2025-12-04T09:31:25.2657228Z inflating: build/bin/c10_hip_HIPTest 2025-12-04T09:31:25.3010154Z inflating: build/bin/vec_test_all_types_DEFAULT 2025-12-04T09:31:25.3362143Z inflating: build/bin/vec_test_all_types_AVX512 2025-12-04T09:31:25.3727900Z inflating: build/bin/vec_test_all_types_AVX2 2025-12-04T09:31:25.3789150Z inflating: build/bin/test_aoti_abi_check 2025-12-04T09:31:25.3819630Z inflating: build/bin/test_vec_half_DEFAULT 2025-12-04T09:31:25.3851870Z inflating: build/bin/test_vec_half_AVX2 2025-12-04T09:31:25.3882385Z inflating: build/bin/test_vec_half_AVX512 2025-12-04T09:31:25.3914296Z inflating: build/bin/BackoffTest 2025-12-04T09:31:25.3947411Z inflating: build/bin/FileStoreTest 2025-12-04T09:31:25.3982535Z inflating: build/bin/TCPStoreTest 2025-12-04T09:31:25.4015123Z inflating: build/bin/HashStoreTest 2025-12-04T09:31:25.4055834Z inflating: build/bin/ProcessGroupGlooTest 2025-12-04T09:31:25.4057319Z inflating: build/bin/example_allreduce 2025-12-04T09:31:25.4059367Z inflating: build/bin/torch_shm_manager 2025-12-04T09:31:25.4102164Z inflating: build/bin/static_runtime_bench 2025-12-04T09:31:25.4260576Z inflating: build/bin/static_runtime_test 2025-12-04T09:31:25.4304213Z inflating: build/bin/Dict_test 2025-12-04T09:31:25.4336546Z inflating: build/bin/Dimname_test 2025-12-04T09:31:25.4375413Z inflating: build/bin/MaybeOwned_test 2025-12-04T09:31:25.4410004Z inflating: build/bin/NamedTensor_test 2025-12-04T09:31:25.4445598Z inflating: build/bin/apply_utils_test 2025-12-04T09:31:25.4482463Z inflating: build/bin/atest 2025-12-04T09:31:25.4528059Z inflating: build/bin/basic 2025-12-04T09:31:25.4562463Z inflating: build/bin/broadcast_test 2025-12-04T09:31:25.4596154Z inflating: build/bin/cpu_allocator_test 2025-12-04T09:31:25.4631239Z inflating: build/bin/cpu_generator_test 2025-12-04T09:31:25.4663119Z inflating: build/bin/cpu_profiling_allocator_test 2025-12-04T09:31:25.4720288Z inflating: build/bin/cpu_rng_test 2025-12-04T09:31:25.4751841Z inflating: build/bin/dlconvertor_test 2025-12-04T09:31:25.4786557Z inflating: build/bin/extension_backend_test 2025-12-04T09:31:25.4820223Z inflating: build/bin/half_test 2025-12-04T09:31:25.4877724Z inflating: build/bin/ivalue_test 2025-12-04T09:31:25.4917091Z inflating: build/bin/lazy_tensor_test 2025-12-04T09:31:25.4949199Z inflating: build/bin/math_kernel_test 2025-12-04T09:31:25.4982367Z inflating: build/bin/memory_format_test 2025-12-04T09:31:25.5014659Z inflating: build/bin/memory_overlapping_test 2025-12-04T09:31:25.5050811Z inflating: build/bin/mobile_memory_cleanup 2025-12-04T09:31:25.5085507Z inflating: build/bin/native_test 2025-12-04T09:31:25.5117576Z inflating: build/bin/operator_name_test 2025-12-04T09:31:25.5148680Z inflating: build/bin/operators_test 2025-12-04T09:31:25.5180237Z inflating: build/bin/packedtensoraccessor_test 2025-12-04T09:31:25.5220649Z inflating: build/bin/pow_test 2025-12-04T09:31:25.5255661Z inflating: build/bin/quantized_test 2025-12-04T09:31:25.5286015Z inflating: build/bin/reduce_ops_test 2025-12-04T09:31:25.5317651Z inflating: build/bin/reportMemoryUsage_test 2025-12-04T09:31:25.5362692Z inflating: build/bin/scalar_tensor_test 2025-12-04T09:31:25.5397071Z inflating: build/bin/scalar_test 2025-12-04T09:31:25.5428507Z inflating: build/bin/StorageUtils_test 2025-12-04T09:31:25.5465590Z inflating: build/bin/stride_properties_test 2025-12-04T09:31:25.5513860Z inflating: build/bin/tensor_iterator_test 2025-12-04T09:31:25.5546611Z inflating: build/bin/test_parallel 2025-12-04T09:31:25.5577570Z inflating: build/bin/thread_init_test 2025-12-04T09:31:25.5613530Z inflating: build/bin/type_ptr_test 2025-12-04T09:31:25.5646904Z inflating: build/bin/type_test 2025-12-04T09:31:25.5679868Z inflating: build/bin/undefined_tensor_test 2025-12-04T09:31:25.5712523Z inflating: build/bin/verify_api_visibility 2025-12-04T09:31:25.5760008Z inflating: build/bin/legacy_vmap_test 2025-12-04T09:31:25.5795700Z inflating: build/bin/weakref_test 2025-12-04T09:31:25.5831653Z inflating: build/bin/wrapdim_test 2025-12-04T09:31:25.5894336Z inflating: build/bin/List_test 2025-12-04T09:31:25.5925488Z inflating: build/bin/xla_tensor_test 2025-12-04T09:31:25.5961185Z inflating: build/bin/IListRef_test 2025-12-04T09:31:25.6032490Z inflating: build/bin/kernel_function_legacy_test 2025-12-04T09:31:25.6072123Z inflating: build/bin/KernelFunction_test 2025-12-04T09:31:25.6134397Z inflating: build/bin/kernel_function_test 2025-12-04T09:31:25.6215493Z inflating: build/bin/kernel_lambda_legacy_test 2025-12-04T09:31:25.6279271Z inflating: build/bin/kernel_lambda_test 2025-12-04T09:31:25.6315499Z inflating: build/bin/kernel_stackbased_test 2025-12-04T09:31:25.6372941Z inflating: build/bin/make_boxed_from_unboxed_functor_test 2025-12-04T09:31:25.6404466Z inflating: build/bin/CppSignature_test 2025-12-04T09:31:25.6434762Z inflating: build/bin/op_allowlist_test 2025-12-04T09:31:25.6625169Z inflating: build/bin/op_registration_test 2025-12-04T09:31:25.6655574Z inflating: build/bin/hip_complex_math_test 2025-12-04T09:31:25.6688997Z inflating: build/bin/backend_fallback_test 2025-12-04T09:31:25.6719058Z inflating: build/bin/hip_complex_test 2025-12-04T09:31:25.6759158Z inflating: build/bin/inline_container_test 2025-12-04T09:31:25.6791462Z inflating: build/bin/hip_apply_test 2025-12-04T09:31:25.6821289Z inflating: build/bin/hip_distributions_test 2025-12-04T09:31:25.6851531Z inflating: build/bin/hip_generator_test 2025-12-04T09:31:25.6881479Z inflating: build/bin/hip_half_test 2025-12-04T09:31:25.6911568Z inflating: build/bin/hip_integer_divider_test 2025-12-04T09:31:25.6953171Z inflating: build/bin/hip_optional_test 2025-12-04T09:31:25.6990992Z inflating: build/bin/hip_packedtensoraccessor_test 2025-12-04T09:31:25.7020815Z inflating: build/bin/hip_vectorized_test 2025-12-04T09:31:25.7052371Z inflating: build/bin/hip_dlconvertor_test 2025-12-04T09:31:25.7690137Z inflating: build/bin/test_jit 2025-12-04T09:31:25.7921127Z inflating: build/bin/test_lazy 2025-12-04T09:31:25.7954867Z inflating: build/bin/test_dist_autograd 2025-12-04T09:31:25.7998169Z inflating: build/bin/test_cpp_rpc 2025-12-04T09:31:25.7999617Z inflating: build/bin/parallel_benchmark 2025-12-04T09:31:25.8704405Z inflating: build/bin/test_api 2025-12-04T09:31:25.8704818Z creating: .additional_ci_files/ 2025-12-04T09:31:25.8740010Z inflating: .additional_ci_files/test-times.json 2025-12-04T09:31:25.8876193Z inflating: .additional_ci_files/test-class-times.json 2025-12-04T09:31:25.8907724Z ##[group]Run rm artifacts.zip 2025-12-04T09:31:25.8907914Z rm artifacts.zip 2025-12-04T09:31:25.8912711Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:31:25.8912870Z env: 2025-12-04T09:31:25.8912977Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:25.8913124Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:25.8913318Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:25.8913498Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:25.8913908Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:25.8914302Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:25.8914429Z AWS_REGION: us-east-1 2025-12-04T09:31:25.8914625Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:25.8914926Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:25.8917308Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:25.8917426Z ##[endgroup] 2025-12-04T09:31:26.0128570Z ##[group]Run df -H 2025-12-04T09:31:26.0128780Z df -H 2025-12-04T09:31:26.0135237Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:31:26.0135468Z env: 2025-12-04T09:31:26.0135626Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:26.0136007Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:26.0136263Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:26.0136508Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:26.0137097Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:26.0137635Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:26.0137798Z AWS_REGION: us-east-1 2025-12-04T09:31:26.0138118Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:26.0138353Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:26.0141444Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:26.0141617Z ##[endgroup] 2025-12-04T09:31:26.0476345Z Filesystem Size Used Avail Use% Mounted on 2025-12-04T09:31:26.0476874Z overlay 16T 738G 15T 5% / 2025-12-04T09:31:26.0477230Z tmpfs 68M 0 68M 0% /dev 2025-12-04T09:31:26.0477662Z /dev/md0 16T 738G 15T 5% /run 2025-12-04T09:31:26.0478012Z shm 68M 4.1k 68M 1% /dev/shm 2025-12-04T09:31:26.0478482Z amdprj2-k8s_2 5.5T 120G 5.4T 3% /home/runner/pytorch-data 2025-12-04T09:31:26.0479009Z tmpfs 3.3T 13k 3.3T 1% /run/secrets/kubernetes.io/serviceaccount 2025-12-04T09:31:26.0479462Z tmpfs 1.7T 0 1.7T 0% /proc/acpi 2025-12-04T09:31:26.0479828Z tmpfs 1.7T 0 1.7T 0% /proc/scsi 2025-12-04T09:31:26.0480193Z tmpfs 1.7T 0 1.7T 0% /sys/firmware 2025-12-04T09:31:26.0480627Z tmpfs 1.7T 0 1.7T 0% /sys/devices/virtual/powercap 2025-12-04T09:31:26.0508152Z Prepare all required actions 2025-12-04T09:31:26.0508400Z Getting action download info 2025-12-04T09:31:26.4575187Z ##[group]Run ./.github/actions/download-td-artifacts 2025-12-04T09:31:26.4575347Z with: 2025-12-04T09:31:26.4575438Z env: 2025-12-04T09:31:26.4575533Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:26.4575688Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:26.4575863Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:26.4576030Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:26.4576414Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:26.4576834Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:26.4576951Z AWS_REGION: us-east-1 2025-12-04T09:31:26.4577125Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:26.4577293Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:26.4579484Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:26.4579588Z ##[endgroup] 2025-12-04T09:31:26.4611570Z ##[group]Run seemethere/download-artifact-s3@v4 2025-12-04T09:31:26.4611703Z with: 2025-12-04T09:31:26.4611793Z name: td_results 2025-12-04T09:31:26.4611902Z s3-bucket: gha-artifacts 2025-12-04T09:31:26.4612009Z region: us-east-1 2025-12-04T09:31:26.4612103Z env: 2025-12-04T09:31:26.4612193Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:26.4612329Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:26.4612507Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:26.4612675Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:26.4613057Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:26.4613429Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:26.4613546Z AWS_REGION: us-east-1 2025-12-04T09:31:26.4613678Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:26.4613826Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:26.4616029Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:26.4616132Z ##[endgroup] 2025-12-04T09:31:26.6913653Z (node:17280) NOTE: We are formalizing our plans to enter AWS SDK for JavaScript (v2) into maintenance mode in 2023. 2025-12-04T09:31:26.6913857Z 2025-12-04T09:31:26.6913946Z Please migrate your code to use AWS SDK for JavaScript (v3). 2025-12-04T09:31:26.6914155Z For more information, check the migration guide at https://a.co/7PzMCcy 2025-12-04T09:31:26.6914370Z (Use `node --trace-warnings ...` to show where the warning was created) 2025-12-04T09:31:26.9608755Z Found 1 objects with prefix pytorch/pytorch/19922849170/td_results/ 2025-12-04T09:31:26.9609417Z Starting download (1/1): /home/runner/_work/pytorch/pytorch/td_results.json 2025-12-04T09:31:27.3651967Z Finished download (1/1): /home/runner/_work/pytorch/pytorch/td_results.json 2025-12-04T09:31:27.3657077Z Artifact download has finished successfully 2025-12-04T09:31:27.3814344Z ##[group]Run mkdir -p .additional_ci_files 2025-12-04T09:31:27.3814514Z mkdir -p .additional_ci_files 2025-12-04T09:31:27.3814685Z mv td_results.json .additional_ci_files/td_results.json || true 2025-12-04T09:31:27.3819395Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:31:27.3819546Z env: 2025-12-04T09:31:27.3819642Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:27.3819779Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:27.3819955Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:27.3820120Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:27.3820659Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:27.3821034Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:27.3821150Z AWS_REGION: us-east-1 2025-12-04T09:31:27.3821345Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:27.3821497Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:27.3823691Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:27.3823800Z ##[endgroup] 2025-12-04T09:31:27.3885078Z ##[group]Run .github/scripts/parse_ref.py 2025-12-04T09:31:27.3885235Z .github/scripts/parse_ref.py 2025-12-04T09:31:27.3890741Z shell: /usr/bin/bash -e {0} 2025-12-04T09:31:27.3890854Z env: 2025-12-04T09:31:27.3890952Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:27.3891093Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:27.3891280Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:27.3891451Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:27.3891842Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:27.3892222Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:27.3892344Z AWS_REGION: us-east-1 2025-12-04T09:31:27.3892490Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:27.3892669Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:27.3894860Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:27.3894973Z ##[endgroup] 2025-12-04T09:31:27.3986148Z Setting output branch=main 2025-12-04T09:31:27.4048725Z Prepare all required actions 2025-12-04T09:31:27.4048959Z Getting action download info 2025-12-04T09:31:27.6132244Z ##[group]Run ./.github/actions/filter-test-configs 2025-12-04T09:31:27.6132399Z with: 2025-12-04T09:31:27.6132606Z github-token: *** 2025-12-04T09:31:27.6135633Z test-matrix: {"include": [{"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}]} 2025-12-04T09:31:27.6139022Z job-name: linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable) 2025-12-04T09:31:27.6139241Z env: 2025-12-04T09:31:27.6139339Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:27.6139483Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:27.6139664Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:27.6139835Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:27.6140222Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:27.6140600Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:27.6140723Z AWS_REGION: us-east-1 2025-12-04T09:31:27.6140934Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:27.6141089Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:27.6143298Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:27.6143407Z ##[endgroup] 2025-12-04T09:31:27.6158845Z ##[group]Run nick-fields/retry@v3.0.0 2025-12-04T09:31:27.6158969Z with: 2025-12-04T09:31:27.6159053Z shell: bash 2025-12-04T09:31:27.6159143Z timeout_minutes: 10 2025-12-04T09:31:27.6159241Z max_attempts: 5 2025-12-04T09:31:27.6159337Z retry_wait_seconds: 30 2025-12-04T09:31:27.6159632Z command: set -eux # PyYAML 6.0 doesn't work with MacOS x86 anymore # This must run on Python-3.7 (AmazonLinux2) so can't use request=3.32.2 python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:31:27.6159934Z polling_interval_seconds: 1 2025-12-04T09:31:27.6160047Z warning_on_retry: true 2025-12-04T09:31:27.6160199Z continue_on_error: false 2025-12-04T09:31:27.6160301Z env: 2025-12-04T09:31:27.6160389Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:27.6160526Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:27.6160701Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:27.6160866Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:27.6161246Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:27.6161612Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:27.6161727Z AWS_REGION: us-east-1 2025-12-04T09:31:27.6161855Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:27.6162004Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:27.6164229Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:27.6164380Z GITHUB_TOKEN: *** 2025-12-04T09:31:27.6164482Z ##[endgroup] 2025-12-04T09:31:27.6558134Z + python3 -m pip install requests==2.27.1 pyyaml==6.0.2 2025-12-04T09:31:27.8064914Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T09:31:27.9037110Z Collecting requests==2.27.1 2025-12-04T09:31:27.9399794Z Downloading requests-2.27.1-py2.py3-none-any.whl (63 kB) 2025-12-04T09:31:27.9499583Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 63.1/63.1 KB 6.4 MB/s eta 0:00:00 2025-12-04T09:31:28.0001749Z Collecting pyyaml==6.0.2 2025-12-04T09:31:28.0082319Z Downloading PyYAML-6.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (751 kB) 2025-12-04T09:31:28.0285597Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 751.2/751.2 KB 39.2 MB/s eta 0:00:00 2025-12-04T09:31:28.0512176Z Collecting idna<4,>=2.5 2025-12-04T09:31:28.0571491Z Downloading idna-3.11-py3-none-any.whl (71 kB) 2025-12-04T09:31:28.0589366Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 71.0/71.0 KB 82.3 MB/s eta 0:00:00 2025-12-04T09:31:28.0788199Z Collecting certifi>=2017.4.17 2025-12-04T09:31:28.0855780Z Downloading certifi-2025.11.12-py3-none-any.whl (159 kB) 2025-12-04T09:31:28.0874762Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 159.4/159.4 KB 179.9 MB/s eta 0:00:00 2025-12-04T09:31:28.1833657Z Collecting charset-normalizer~=2.0.0 2025-12-04T09:31:28.1895169Z Downloading charset_normalizer-2.0.12-py3-none-any.whl (39 kB) 2025-12-04T09:31:28.2204026Z Collecting urllib3<1.27,>=1.21.1 2025-12-04T09:31:28.2265687Z Downloading urllib3-1.26.20-py2.py3-none-any.whl (144 kB) 2025-12-04T09:31:28.2285499Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 144.2/144.2 KB 164.7 MB/s eta 0:00:00 2025-12-04T09:31:28.2909579Z Installing collected packages: urllib3, pyyaml, idna, charset-normalizer, certifi, requests 2025-12-04T09:31:28.3992112Z WARNING: The script normalizer is installed in '/home/runner/.local/bin' which is not on PATH. 2025-12-04T09:31:28.3992514Z Consider adding this directory to PATH or, if you prefer to suppress this warning, use --no-warn-script-location. 2025-12-04T09:31:28.4176938Z Successfully installed certifi-2025.11.12 charset-normalizer-2.0.12 idna-3.11 pyyaml-6.0.2 requests-2.27.1 urllib3-1.26.20 2025-12-04T09:31:28.6564556Z Command completed after 1 attempt(s). 2025-12-04T09:31:28.6609480Z ##[group]Run set -x 2025-12-04T09:31:28.6609623Z set -x 2025-12-04T09:31:28.6609719Z  2025-12-04T09:31:28.6609874Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:31:28.6610059Z # in runner workspace 2025-12-04T09:31:28.6610217Z python3 "${GITHUB_ACTION_PATH}/../../scripts/parse_ref.py" 2025-12-04T09:31:28.6614863Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:31:28.6615020Z env: 2025-12-04T09:31:28.6615118Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:28.6615257Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:28.6615436Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:28.6615748Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:28.6616149Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:28.6616520Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:28.6616637Z AWS_REGION: us-east-1 2025-12-04T09:31:28.6616856Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:28.6617011Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:28.6619194Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:28.6619306Z ##[endgroup] 2025-12-04T09:31:28.6643862Z + python3 /home/runner/_work/pytorch/pytorch/./.github/actions/filter-test-configs/../../scripts/parse_ref.py 2025-12-04T09:31:28.6728626Z Setting output branch=main 2025-12-04T09:31:28.6758493Z ##[group]Run echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:31:28.6758690Z echo "Workflow: ${GITHUB_WORKFLOW}" 2025-12-04T09:31:28.6758837Z echo "Job name: ${JOB_NAME}" 2025-12-04T09:31:28.6758985Z  2025-12-04T09:31:28.6759139Z # Use relative path here as this could be checked out anywhere, not necessarily 2025-12-04T09:31:28.6759337Z # in runner workspace 2025-12-04T09:31:28.6759511Z python3 "${GITHUB_ACTION_PATH}/../../scripts/filter_test_configs.py" \ 2025-12-04T09:31:28.6759701Z  --workflow "${GITHUB_WORKFLOW}" \ 2025-12-04T09:31:28.6759839Z  --job-name "${JOB_NAME}" \ 2025-12-04T09:31:28.6763018Z  --test-matrix "{"include": [{"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}]}" \ 2025-12-04T09:31:28.6766236Z  --selected-test-configs "" \ 2025-12-04T09:31:28.6766375Z  --pr-number "${PR_NUMBER}" \ 2025-12-04T09:31:28.6766507Z  --tag "${TAG}" \ 2025-12-04T09:31:28.6766632Z  --event-name "${EVENT_NAME}" \ 2025-12-04T09:31:28.6766836Z  --schedule "${SCHEDULE}" \ 2025-12-04T09:31:28.6766958Z  --branch "${HEAD_BRANCH}" 2025-12-04T09:31:28.6771152Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:31:28.6771298Z env: 2025-12-04T09:31:28.6771392Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:28.6771526Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:28.6771704Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:28.6771869Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:28.6772251Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:28.6772622Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:28.6772737Z AWS_REGION: us-east-1 2025-12-04T09:31:28.6772903Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:28.6773053Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:28.6775239Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:28.6775425Z GITHUB_TOKEN: *** 2025-12-04T09:31:28.6775624Z JOB_NAME: linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable) 2025-12-04T09:31:28.6775834Z PR_NUMBER: 2025-12-04T09:31:28.6775925Z TAG: 2025-12-04T09:31:28.6776012Z EVENT_NAME: schedule 2025-12-04T09:31:28.6776114Z SCHEDULE: 29 8 * * * 2025-12-04T09:31:28.6776209Z HEAD_BRANCH: main 2025-12-04T09:31:28.6776306Z ##[endgroup] 2025-12-04T09:31:28.6794700Z Workflow: trunk-rocm-mi300 2025-12-04T09:31:28.6794935Z Job name: linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable) 2025-12-04T09:31:29.2768019Z INFO:root:Issue https://github.com/pytorch/pytorch/issues/167616 created by jithunnair-amd has unstable all the test jobs for trunk-rocm-mi300 / linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable) 2025-12-04T09:31:29.7740935Z Setting output keep-going=True 2025-12-04T09:31:29.7741268Z Setting output ci-verbose-test-logs=False 2025-12-04T09:31:29.7741547Z Setting output ci-test-showlocals=False 2025-12-04T09:31:29.7741816Z Setting output ci-no-test-timeout=False 2025-12-04T09:31:29.7742055Z Setting output ci-no-td=False 2025-12-04T09:31:29.7742278Z Setting output ci-td-distributed=False 2025-12-04T09:31:29.7742539Z Setting output is-unstable=True 2025-12-04T09:31:29.7742765Z Setting output reenabled-issues= 2025-12-04T09:31:29.7755689Z Setting output test-matrix={"include": [{"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}]} 2025-12-04T09:31:29.7764956Z Setting output is-test-matrix-empty=False 2025-12-04T09:31:29.7829947Z ##[group]Run echo "Filtered matrix:" 2025-12-04T09:31:29.7830191Z echo "Filtered matrix:" 2025-12-04T09:31:29.7838356Z echo "{"include": [{"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 1, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 2, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 3, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 4, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 5, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "default", "shard": 6, "num_shards": 6, "runner": "linux.rocm.gpu.gfx942.1.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 1, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 2, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "mem_leak_check": "mem_leak_check", "unstable": "unstable", "rerun_disabled_tests": "rerun_disabled_tests"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable", "mem_leak_check": "mem_leak_check"}, {"config": "distributed", "shard": 3, "num_shards": 3, "runner": "linux.rocm.gpu.gfx942.4.b", "rerun_disabled_tests": "rerun_disabled_tests", "unstable": "unstable"}]}" 2025-12-04T09:31:29.7845633Z  2025-12-04T09:31:29.7845722Z echo 2025-12-04T09:31:29.7845841Z echo "Is the current job unstable? True" 2025-12-04T09:31:29.7845973Z  2025-12-04T09:31:29.7846060Z echo 2025-12-04T09:31:29.7846168Z echo "Is keep-going label set? True" 2025-12-04T09:31:29.7846294Z  2025-12-04T09:31:29.7846378Z echo 2025-12-04T09:31:29.7846478Z echo "Reenabled issues? " 2025-12-04T09:31:29.7851391Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:31:29.7851540Z env: 2025-12-04T09:31:29.7851639Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:29.7851776Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:29.7851953Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:29.7852122Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:29.7852515Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:29.7852884Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:29.7853002Z AWS_REGION: us-east-1 2025-12-04T09:31:29.7853185Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:29.7853337Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:29.7855527Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:29.7855640Z ##[endgroup] 2025-12-04T09:31:29.7878120Z Filtered matrix: 2025-12-04T09:31:29.7886820Z {include: [{config: default, shard: 1, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable}, {config: default, shard: 1, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 1, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: default, shard: 1, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: default, shard: 2, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable}, {config: default, shard: 2, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 2, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: default, shard: 2, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: default, shard: 3, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable}, {config: default, shard: 3, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 3, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: default, shard: 3, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: default, shard: 4, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable}, {config: default, shard: 4, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 4, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: default, shard: 4, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: default, shard: 5, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable}, {config: default, shard: 5, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 5, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: default, shard: 5, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: default, shard: 6, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable}, {config: default, shard: 6, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: default, shard: 6, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: default, shard: 6, num_shards: 6, runner: linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: distributed, shard: 1, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, mem_leak_check: mem_leak_check, unstable: unstable}, {config: distributed, shard: 1, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: distributed, shard: 1, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: distributed, shard: 1, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: distributed, shard: 2, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, mem_leak_check: mem_leak_check, unstable: unstable}, {config: distributed, shard: 2, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: distributed, shard: 2, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: distributed, shard: 2, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}, {config: distributed, shard: 3, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, mem_leak_check: mem_leak_check, unstable: unstable}, {config: distributed, shard: 3, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, mem_leak_check: mem_leak_check, unstable: unstable, rerun_disabled_tests: rerun_disabled_tests}, {config: distributed, shard: 3, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable, mem_leak_check: mem_leak_check}, {config: distributed, shard: 3, num_shards: 3, runner: linux.rocm.gpu.gfx942.4.b, rerun_disabled_tests: rerun_disabled_tests, unstable: unstable}]} 2025-12-04T09:31:29.7893718Z 2025-12-04T09:31:29.7894383Z Is the current job unstable? True 2025-12-04T09:31:29.7894465Z 2025-12-04T09:31:29.7894516Z Is keep-going label set? True 2025-12-04T09:31:29.7894594Z 2025-12-04T09:31:29.7894635Z Reenabled issues? 2025-12-04T09:31:29.7924652Z ##[group]Run echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:31:29.7924875Z echo "timeout=$((JOB_TIMEOUT-30))" >> "${GITHUB_OUTPUT}" 2025-12-04T09:31:29.7929916Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:31:29.7930063Z env: 2025-12-04T09:31:29.7930155Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:29.7930287Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:29.7930461Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:29.7930625Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:29.7931083Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:29.7931452Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:29.7931570Z AWS_REGION: us-east-1 2025-12-04T09:31:29.7931758Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:29.7931940Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:29.7934158Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:29.7934261Z JOB_TIMEOUT: 300 2025-12-04T09:31:29.7934363Z ##[endgroup] 2025-12-04T09:31:29.7974787Z ##[group]Run env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:31:29.7975019Z env | grep '^GITHUB' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:31:29.7975208Z env | grep '^CI' >> "/tmp/github_env_${GITHUB_RUN_ID}" 2025-12-04T09:31:29.7980130Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T09:31:29.7980304Z env: 2025-12-04T09:31:29.7980410Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:29.7980554Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:29.7980743Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:29.7980920Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:29.7981319Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:29.7981721Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:29.7981841Z AWS_REGION: us-east-1 2025-12-04T09:31:29.7982026Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:29.7982183Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:29.7984367Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:29.7984483Z ##[endgroup] 2025-12-04T09:31:29.8079489Z ##[group]Run set -x 2025-12-04T09:31:29.8079674Z set -x 2025-12-04T09:31:29.8079773Z  2025-12-04T09:31:29.8079888Z if [[ $TEST_CONFIG == 'multigpu' ]]; then 2025-12-04T09:31:29.8080052Z  TEST_COMMAND=.ci/pytorch/multigpu-test.sh 2025-12-04T09:31:29.8080213Z elif [[ $BUILD_ENVIRONMENT == *onnx* ]]; then 2025-12-04T09:31:29.8080361Z  TEST_COMMAND=.ci/caffe2/test.sh 2025-12-04T09:31:29.8080487Z else 2025-12-04T09:31:29.8080597Z  TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:31:29.8080722Z fi 2025-12-04T09:31:29.8080814Z  2025-12-04T09:31:29.8080955Z # detached container should get cleaned up by teardown_ec2_linux 2025-12-04T09:31:29.8081163Z # TODO: Stop building test binaries as part of the build phase 2025-12-04T09:31:29.8081348Z # Used for GPU_FLAG since that doesn't play nice 2025-12-04T09:31:29.8081521Z # shellcheck disable=SC2086,SC2090 2025-12-04T09:31:29.8081662Z container_name=$(docker run \ 2025-12-04T09:31:29.8081799Z  ${GPU_FLAG:-} \ 2025-12-04T09:31:29.8081921Z  -e BUILD_ENVIRONMENT \ 2025-12-04T09:31:29.8082048Z  -e PR_NUMBER \ 2025-12-04T09:31:29.8082167Z  -e GITHUB_ACTIONS \ 2025-12-04T09:31:29.8082290Z  -e GITHUB_REPOSITORY \ 2025-12-04T09:31:29.8082415Z  -e GITHUB_WORKFLOW \ 2025-12-04T09:31:29.8082536Z  -e GITHUB_JOB \ 2025-12-04T09:31:29.8082653Z  -e GITHUB_RUN_ID \ 2025-12-04T09:31:29.8082768Z  -e GITHUB_RUN_NUMBER \ 2025-12-04T09:31:29.8082895Z  -e GITHUB_RUN_ATTEMPT \ 2025-12-04T09:31:29.8083020Z  -e JOB_ID \ 2025-12-04T09:31:29.8083133Z  -e JOB_NAME \ 2025-12-04T09:31:29.8083246Z  -e BASE_SHA \ 2025-12-04T09:31:29.8083356Z  -e BRANCH \ 2025-12-04T09:31:29.8083462Z  -e SHA1 \ 2025-12-04T09:31:29.8083576Z  -e AWS_DEFAULT_REGION \ 2025-12-04T09:31:29.8083701Z  -e IN_WHEEL_TEST \ 2025-12-04T09:31:29.8083942Z  -e SHARD_NUMBER \ 2025-12-04T09:31:29.8084062Z  -e TEST_CONFIG \ 2025-12-04T09:31:29.8084184Z  -e NUM_TEST_SHARDS \ 2025-12-04T09:31:29.8084309Z  -e REENABLED_ISSUES \ 2025-12-04T09:31:29.8084437Z  -e CONTINUE_THROUGH_ERROR \ 2025-12-04T09:31:29.8084568Z  -e VERBOSE_TEST_LOGS \ 2025-12-04T09:31:29.8084695Z  -e TEST_SHOWLOCALS \ 2025-12-04T09:31:29.8084815Z  -e NO_TEST_TIMEOUT \ 2025-12-04T09:31:29.8084932Z  -e NO_TD \ 2025-12-04T09:31:29.8085056Z  -e MAX_JOBS="$(nproc --ignore=2)" \ 2025-12-04T09:31:29.8085208Z  -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK \ 2025-12-04T09:31:29.8085358Z  -e PYTORCH_TEST_RERUN_DISABLED_TESTS \ 2025-12-04T09:31:29.8085498Z  -e TESTS_TO_INCLUDE \ 2025-12-04T09:31:29.8085624Z  -e HUGGING_FACE_HUB_TOKEN \ 2025-12-04T09:31:29.8085761Z  -e DASHBOARD_TAG \ 2025-12-04T09:31:29.8085924Z  --env-file="${RUNNER_TEMP}/github_env_${GITHUB_RUN_ID}" \ 2025-12-04T09:31:29.8086093Z  --ulimit stack=10485760:83886080 \ 2025-12-04T09:31:29.8086229Z  --ulimit core=0 \ 2025-12-04T09:31:29.8086370Z  --env-file="/tmp/github_env_${GITHUB_RUN_ID}" \ 2025-12-04T09:31:29.8086536Z  --security-opt seccomp=unconfined \ 2025-12-04T09:31:29.8086682Z  --cap-add=SYS_PTRACE \ 2025-12-04T09:31:29.8086862Z  --shm-size="8g" \ 2025-12-04T09:31:29.8086983Z  --tty \ 2025-12-04T09:31:29.8087096Z  --detach \ 2025-12-04T09:31:29.8087216Z  --name="${container_name}" \ 2025-12-04T09:31:29.8087351Z  --user jenkins \ 2025-12-04T09:31:29.8087502Z  -v "${GITHUB_WORKSPACE}:/var/lib/jenkins/workspace" \ 2025-12-04T09:31:29.8087666Z  -w /var/lib/jenkins/workspace \ 2025-12-04T09:31:29.8087861Z  "${DOCKER_IMAGE}" 2025-12-04T09:31:29.8087978Z ) 2025-12-04T09:31:29.8088105Z # save container name for later step 2025-12-04T09:31:29.8088281Z echo "CONTAINER_NAME=${container_name}" >> "$GITHUB_ENV" 2025-12-04T09:31:29.8088566Z # jenkins user does not have write permission to mounted workspace; work-around by copying within container to jenkins home 2025-12-04T09:31:29.8088925Z docker exec -t "${container_name}" sh -c "cd .. && cp -R workspace pytorch && cd pytorch && pip install dist/*.whl && ${TEST_COMMAND}" 2025-12-04T09:31:29.8093336Z shell: /usr/bin/bash -e {0} 2025-12-04T09:31:29.8093465Z env: 2025-12-04T09:31:29.8093574Z GIT_DEFAULT_BRANCH: main 2025-12-04T09:31:29.8093724Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T09:31:29.8093916Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T09:31:29.8094097Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T09:31:29.8094502Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T09:31:29.8094887Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T09:31:29.8095020Z AWS_REGION: us-east-1 2025-12-04T09:31:29.8095199Z AWS_ACCESS_KEY_ID: *** 2025-12-04T09:31:29.8095366Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T09:31:29.8097629Z AWS_SESSION_TOKEN: *** 2025-12-04T09:31:29.8097762Z BUILD_ENVIRONMENT: linux-jammy-rocm-py3.10 2025-12-04T09:31:29.8097905Z PR_NUMBER: 2025-12-04T09:31:29.8098023Z GITHUB_REPOSITORY: pytorch/pytorch 2025-12-04T09:31:29.8098165Z GITHUB_WORKFLOW: trunk-rocm-mi300 2025-12-04T09:31:29.8098298Z GITHUB_JOB: test 2025-12-04T09:31:29.8098411Z GITHUB_RUN_ID: 19922849170 2025-12-04T09:31:29.8098528Z GITHUB_RUN_NUMBER: 689 2025-12-04T09:31:29.8098647Z GITHUB_RUN_ATTEMPT: 1 2025-12-04T09:31:29.8098764Z JOB_ID: 57116213149 2025-12-04T09:31:29.8098980Z JOB_NAME: linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable) 2025-12-04T09:31:29.8099258Z BRANCH: main 2025-12-04T09:31:29.8099382Z SHA1: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:31:29.8099546Z BASE_SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:31:29.8099694Z TEST_CONFIG: default 2025-12-04T09:31:29.8099806Z SHARD_NUMBER: 4 2025-12-04T09:31:29.8099913Z NUM_TEST_SHARDS: 6 2025-12-04T09:31:29.8100028Z REENABLED_ISSUES: 2025-12-04T09:31:29.8100147Z CONTINUE_THROUGH_ERROR: True 2025-12-04T09:31:29.8100277Z VERBOSE_TEST_LOGS: False 2025-12-04T09:31:29.8100400Z TEST_SHOWLOCALS: False 2025-12-04T09:31:29.8100530Z NO_TEST_TIMEOUT: False 2025-12-04T09:31:29.8100644Z NO_TD: False 2025-12-04T09:31:29.8100919Z DOCKER_IMAGE: 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:31:29.8101228Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK: 0 2025-12-04T09:31:29.8101376Z PYTORCH_TEST_RERUN_DISABLED_TESTS: 1 2025-12-04T09:31:29.8101517Z TESTS_TO_INCLUDE: 2025-12-04T09:31:29.8101629Z DASHBOARD_TAG: 2025-12-04T09:31:29.8101787Z HUGGING_FACE_HUB_TOKEN: *** 2025-12-04T09:31:29.8101916Z ##[endgroup] 2025-12-04T09:31:29.8118198Z + [[ default == \m\u\l\t\i\g\p\u ]] 2025-12-04T09:31:29.8118340Z + [[ linux-jammy-rocm-py3.10 == *onnx* ]] 2025-12-04T09:31:29.8118487Z + TEST_COMMAND=.ci/pytorch/test.sh 2025-12-04T09:31:29.8127410Z +++ nproc --ignore=2 2025-12-04T09:31:29.8137888Z ++ docker run --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host -e BUILD_ENVIRONMENT -e PR_NUMBER -e GITHUB_ACTIONS -e GITHUB_REPOSITORY -e GITHUB_WORKFLOW -e GITHUB_JOB -e GITHUB_RUN_ID -e GITHUB_RUN_NUMBER -e GITHUB_RUN_ATTEMPT -e JOB_ID -e JOB_NAME -e BASE_SHA -e BRANCH -e SHA1 -e AWS_DEFAULT_REGION -e IN_WHEEL_TEST -e SHARD_NUMBER -e TEST_CONFIG -e NUM_TEST_SHARDS -e REENABLED_ISSUES -e CONTINUE_THROUGH_ERROR -e VERBOSE_TEST_LOGS -e TEST_SHOWLOCALS -e NO_TEST_TIMEOUT -e NO_TD -e MAX_JOBS=126 -e PYTORCH_TEST_CUDA_MEM_LEAK_CHECK -e PYTORCH_TEST_RERUN_DISABLED_TESTS -e TESTS_TO_INCLUDE -e HUGGING_FACE_HUB_TOKEN -e DASHBOARD_TAG --env-file=/home/runner/_work/_temp/github_env_19922849170 --ulimit stack=10485760:83886080 --ulimit core=0 --env-file=/tmp/github_env_19922849170 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --shm-size=8g --tty --detach --name= --user jenkins -v /home/runner/_work/pytorch/pytorch:/var/lib/jenkins/workspace -w /var/lib/jenkins/workspace 308535385114.dkr.ecr.us-east-1.amazonaws.com/pytorch/ci-image:pytorch-linux-jammy-rocm-n-py3-f0cd68561080d537ef3d3d6f81b25a6416ad600a 2025-12-04T09:31:30.0748735Z + container_name=eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T09:31:30.0749261Z + echo CONTAINER_NAME=eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T09:31:30.0751275Z + docker exec -t eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa sh -c 'cd .. && cp -R workspace pytorch && cd pytorch && pip install dist/*.whl && .ci/pytorch/test.sh' 2025-12-04T09:31:33.7316260Z Processing ./dist/torch-2.10.0a0+gitffd9b0f-cp310-cp310-linux_x86_64.whl 2025-12-04T09:31:34.2695572Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f) (3.18.0) 2025-12-04T09:31:34.2696340Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f) (4.12.2) 2025-12-04T09:31:34.2697484Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f) (1.13.3) 2025-12-04T09:31:34.2701685Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f) (2.8.8) 2025-12-04T09:31:34.2703777Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f) (3.1.6) 2025-12-04T09:31:34.2704318Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch==2.10.0a0+gitffd9b0f) (2025.10.0) 2025-12-04T09:31:34.2866134Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch==2.10.0a0+gitffd9b0f) (1.3.0) 2025-12-04T09:31:34.2890146Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch==2.10.0a0+gitffd9b0f) (3.0.3) 2025-12-04T09:31:34.4879338Z Installing collected packages: torch 2025-12-04T09:31:40.2705053Z Successfully installed torch-2.10.0a0+gitffd9b0f 2025-12-04T09:31:40.3125279Z + export TERM=vt100 2025-12-04T09:31:40.3125455Z + TERM=vt100 2025-12-04T09:31:40.3129254Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:31:40.3140761Z + source .ci/pytorch/common.sh 2025-12-04T09:31:40.3148293Z +++ dirname .ci/pytorch/common.sh 2025-12-04T09:31:40.3160797Z ++ source .ci/pytorch/common_utils.sh 2025-12-04T09:31:40.3162514Z +++ declare -f -t trap_add 2025-12-04T09:31:40.3166639Z ++ set -ex -o pipefail 2025-12-04T09:31:40.3166937Z ++ [[ linux-jammy-rocm-py3.10 == *rocm* ]] 2025-12-04T09:31:40.3167134Z ++ unset HIP_PLATFORM 2025-12-04T09:31:40.3167290Z ++ export PYTORCH_TEST_WITH_ROCM=1 2025-12-04T09:31:40.3167476Z ++ PYTORCH_TEST_WITH_ROCM=1 2025-12-04T09:31:40.3167634Z ++ BUILD_TEST_LIBTORCH=0 2025-12-04T09:31:40.3171170Z ++ dirname .ci/pytorch/test.sh 2025-12-04T09:31:40.3176134Z + source .ci/pytorch/common-build.sh 2025-12-04T09:31:40.3177979Z ++ [[ linux-jammy-rocm-py3.10 != *win-* ]] 2025-12-04T09:31:40.3187894Z ++++ dirname .ci/pytorch/common-build.sh 2025-12-04T09:31:40.3197732Z +++ cd .ci/pytorch 2025-12-04T09:31:40.3198071Z +++ pwd -P 2025-12-04T09:31:40.3202678Z ++ script_dir=/var/lib/jenkins/pytorch/.ci/pytorch 2025-12-04T09:31:40.3202976Z ++ [[ linux-jammy-rocm-py3.10 == *-pch* ]] 2025-12-04T09:31:40.3203206Z ++ which sccache 2025-12-04T09:31:40.3215941Z ++ [[ -z '' ]] 2025-12-04T09:31:40.3216107Z ++ unset SCCACHE_BUCKET 2025-12-04T09:31:40.3216264Z ++ unset SCCACHE_REGION 2025-12-04T09:31:40.3216421Z ++ sccache --stop-server 2025-12-04T09:31:40.3238425Z ++ true 2025-12-04T09:31:40.3238572Z ++ rm -f /var/lib/jenkins/sccache_error.log 2025-12-04T09:31:40.3247445Z ++ trap_add sccache_epilogue EXIT 2025-12-04T09:31:40.3247632Z ++ trap_add_cmd=sccache_epilogue 2025-12-04T09:31:40.3247791Z ++ shift 2025-12-04T09:31:40.3247927Z ++ for trap_add_name in "$@" 2025-12-04T09:31:40.3254599Z ++++ trap -p EXIT 2025-12-04T09:31:40.3257070Z +++ eval 'extract_trap_cmd ' 2025-12-04T09:31:40.3257230Z ++++ extract_trap_cmd 2025-12-04T09:31:40.3257364Z ++++ printf '%s\n' '' 2025-12-04T09:31:40.3257531Z +++ printf '%s\n' sccache_epilogue 2025-12-04T09:31:40.3259425Z ++ trap -- ' 2025-12-04T09:31:40.3259569Z sccache_epilogue' EXIT 2025-12-04T09:31:40.3259705Z ++ [[ -n '' ]] 2025-12-04T09:31:40.3259859Z ++ [[ linux-jammy-rocm-py3.10 == *rocm* ]] 2025-12-04T09:31:40.3260062Z ++ SCCACHE_ERROR_LOG=/var/lib/jenkins/sccache_error.log 2025-12-04T09:31:40.3260256Z ++ SCCACHE_IDLE_TIMEOUT=0 2025-12-04T09:31:40.3260407Z ++ sccache --start-server 2025-12-04T09:31:40.3281558Z sccache: Starting the server... 2025-12-04T09:31:40.3565103Z sccache: Listening on address 127.0.0.1:4226 2025-12-04T09:31:40.3574781Z ++ sccache --zero-stats 2025-12-04T09:31:40.3594025Z Statistics zeroed. 2025-12-04T09:31:40.3597701Z ++ which ccache 2025-12-04T09:31:40.3609723Z + [[ linux-jammy-rocm-py3.10 != *rocm* ]] 2025-12-04T09:31:40.3611137Z + [[ linux-jammy-rocm-py3.10 == *cuda* ]] 2025-12-04T09:31:40.3611271Z + echo 'Environment variables:' 2025-12-04T09:31:40.3611403Z Environment variables: 2025-12-04T09:31:40.3611507Z + env 2025-12-04T09:31:40.3616634Z GITHUB_WORKSPACE=/home/runner/_work/pytorch/pytorch 2025-12-04T09:31:40.3616861Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:31:40.3616994Z BUILD_ENVIRONMENT=linux-jammy-rocm-py3.10 2025-12-04T09:31:40.3618714Z HOSTNAME=linux.rocm.gpu.gfx942.1.b-gwk9b-runner-68hm6 2025-12-04T09:31:40.3618956Z GITHUB_PATH=/home/runner/_work/_temp/_runner_file_commands/add_path_6a3264ba-5c74-4e58-a72b-bbfb124cda02 2025-12-04T09:31:40.3619165Z GITHUB_ACTION=__run_2 2025-12-04T09:31:40.3619275Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-12-04T09:31:40.3619398Z GITHUB_RUN_NUMBER=689 2025-12-04T09:31:40.3619502Z TEST_CONFIG=default 2025-12-04T09:31:40.3619640Z RUNNER_NAME=linux.rocm.gpu.gfx942.1.b-gwk9b-runner-68hm6 2025-12-04T09:31:40.3619801Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:31:40.3619932Z AWS_DEFAULT_REGION=us-east-1 2025-12-04T09:31:40.3620070Z RUNNER_ARTIFACT_DIR=/home/runner/_work/_temp/artifacts 2025-12-04T09:31:40.3620280Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-12-04T09:31:40.3620406Z GITHUB_REF_TYPE=branch 2025-12-04T09:31:40.3620531Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:31:40.3620826Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:31:40.3623598Z *** 2025-12-04T09:31:40.3623698Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:31:40.3623816Z GITHUB_ACTIONS=true 2025-12-04T09:31:40.3623933Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:31:40.3624085Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:31:40.3624305Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/trunk-rocm-mi300.yml@refs/heads/main 2025-12-04T09:31:40.3624496Z UCC_HOME=/usr 2025-12-04T09:31:40.3624598Z RUNNER_ENVIRONMENT=self-hosted 2025-12-04T09:31:40.3624715Z VERBOSE_TEST_LOGS=False 2025-12-04T09:31:40.3624824Z GITHUB_REF=refs/heads/main 2025-12-04T09:31:40.3624929Z RUNNER_OS=Linux 2025-12-04T09:31:40.3625022Z SHARD_NUMBER=4 2025-12-04T09:31:40.3625119Z GITHUB_REF_PROTECTED=true 2025-12-04T09:31:40.3625231Z RUNNER_MANUALLY_TRAP_SIG=1 2025-12-04T09:31:40.3625341Z HOME=/var/lib/jenkins 2025-12-04T09:31:40.3625466Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:31:40.3625690Z PYTORCH_TEST_RERUN_DISABLED_TESTS=1 2025-12-04T09:31:40.3625827Z RUNNER_DOCS_DIR=/home/runner/_work/_temp/docs 2025-12-04T09:31:40.3625952Z LANG=C.UTF-8 2025-12-04T09:31:40.3626067Z UCX_COMMIT=29831d319e6be55cb8c768ca61de335c934ca39e 2025-12-04T09:31:40.3626206Z PYTORCH_TEST_WITH_ROCM=1 2025-12-04T09:31:40.3626348Z RUNNER_TRACKING_ID=github_c5d5cb89-1e45-4146-be76-5d780902345a 2025-12-04T09:31:40.3626496Z RUNNER_ARCH=X64 2025-12-04T09:31:40.3626599Z RUNNER_TEMP=/home/runner/_work/_temp 2025-12-04T09:31:40.3626717Z NUM_TEST_SHARDS=6 2025-12-04T09:31:40.3626900Z UCX_HOME=/usr 2025-12-04T09:31:40.3627093Z GITHUB_STATE=/home/runner/_work/_temp/_runner_file_commands/save_state_6a3264ba-5c74-4e58-a72b-bbfb124cda02 2025-12-04T09:31:40.3627409Z JOB_NAME=linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable) 2025-12-04T09:31:40.3627626Z MAGMA_HOME=/opt/rocm/magma 2025-12-04T09:31:40.3627825Z GITHUB_ENV=/home/runner/_work/_temp/_runner_file_commands/set_env_6a3264ba-5c74-4e58-a72b-bbfb124cda02 2025-12-04T09:31:40.3628071Z GITHUB_EVENT_PATH=/home/runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:31:40.3628237Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:31:40.3628395Z GITHUB_ACTIONS_RUNNER_EXTRA_USER_AGENT=actions-runner-controller/0.12.1 2025-12-04T09:31:40.3628561Z DASHBOARD_TAG= 2025-12-04T09:31:40.3628660Z GITHUB_RUN_ID=19922849170 2025-12-04T09:31:40.3628877Z GITHUB_STEP_SUMMARY=/home/runner/_work/_temp/_runner_file_commands/step_summary_6a3264ba-5c74-4e58-a72b-bbfb124cda02 2025-12-04T09:31:40.3629106Z GITHUB_ACTOR=pytorchmergebot 2025-12-04T09:31:40.3629218Z PR_NUMBER= 2025-12-04T09:31:40.3629311Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:31:40.3629421Z ANACONDA_PYTHON_VERSION=3.10 2025-12-04T09:31:40.3629557Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:31:40.3629691Z TERM=vt100 2025-12-04T09:31:40.3629783Z INSTALLED_VISION=yes 2025-12-04T09:31:40.3629886Z BRANCH=main 2025-12-04T09:31:40.3629985Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:31:40.3630099Z TESTS_TO_INCLUDE= 2025-12-04T09:31:40.3630262Z GITHUB_ACTION_PATH=/home/runner/_work/pytorch/pytorch/./.github/actions/setup-rocm 2025-12-04T09:31:40.3630550Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:31:40.3630688Z PYTORCH_ROCM_ARCH=gfx90a;gfx942;gfx950;gfx1100 2025-12-04T09:31:40.3630841Z UCC_COMMIT=9f4b242cbbd8b1462cbc732eb29316cdfa124b77 2025-12-04T09:31:40.3630977Z REENABLED_ISSUES= 2025-12-04T09:31:40.3631071Z SHLVL=1 2025-12-04T09:31:40.3631159Z MAX_JOBS=126 2025-12-04T09:31:40.3631288Z RUNNER_TEST_RESULTS_DIR=/home/runner/_work/_temp/test-results 2025-12-04T09:31:40.3631441Z GITHUB_ACTOR_ID=97764156 2025-12-04T09:31:40.3631558Z RUNNER_TOOL_CACHE=/home/runner/_work/_tool 2025-12-04T09:31:40.3631717Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:31:40.3631885Z GITHUB_REF_NAME=main 2025-12-04T09:31:40.3631987Z ROCM_PATH=/opt/rocm 2025-12-04T09:31:40.3632085Z GITHUB_JOB=test 2025-12-04T09:31:40.3632182Z NO_TEST_TIMEOUT=False 2025-12-04T09:31:40.3632299Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:31:40.3632416Z LC_ALL=C.UTF-8 2025-12-04T09:31:40.3632521Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:31:40.3632644Z RUNNER_WORKSPACE=/home/runner/_work/pytorch 2025-12-04T09:31:40.3632773Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:31:40.3632885Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:31:40.3633238Z PATH=/opt/cache/bin:/opt/rocm/llvm/bin:/opt/rocm/opencl/bin:/opt/rocm/hip/bin:/opt/rocm/hcc/bin:/opt/rocm/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:31:40.3633588Z GITHUB_BASE_REF= 2025-12-04T09:31:40.3633684Z CI=true 2025-12-04T09:31:40.3633779Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:31:40.3633890Z JOB_ID=57116213149 2025-12-04T09:31:40.3633985Z GITHUB_HEAD_REF= 2025-12-04T09:31:40.3634080Z GITHUB_ACTION_REF= 2025-12-04T09:31:40.3634178Z TEST_SHOWLOCALS=False 2025-12-04T09:31:40.3634290Z GITHUB_WORKFLOW=trunk-rocm-mi300 2025-12-04T09:31:40.3634414Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:31:40.3634668Z GITHUB_OUTPUT=/home/runner/_work/_temp/_runner_file_commands/set_output_6a3264ba-5c74-4e58-a72b-bbfb124cda02 2025-12-04T09:31:40.3634882Z NO_TD=False 2025-12-04T09:31:40.3634977Z OLDPWD=/var/lib/jenkins 2025-12-04T09:31:40.3635081Z _=/usr/bin/env 2025-12-04T09:31:40.3635224Z ++ python -c 'import site; print(site.getsitepackages()[0])' 2025-12-04T09:31:40.3687434Z + TORCH_INSTALL_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch 2025-12-04T09:31:40.3688769Z + TORCH_BIN_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/bin 2025-12-04T09:31:40.3689359Z + TORCH_LIB_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib 2025-12-04T09:31:40.3689844Z + TORCH_TEST_DIR=/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/test 2025-12-04T09:31:40.3690211Z + BUILD_DIR=build 2025-12-04T09:31:40.3690446Z + BUILD_RENAMED_DIR=build_renamed 2025-12-04T09:31:40.3690709Z + BUILD_BIN_DIR=build/bin 2025-12-04T09:31:40.3690937Z + SHARD_NUMBER=4 2025-12-04T09:31:40.3691208Z + NUM_TEST_SHARDS=6 2025-12-04T09:31:40.3691441Z + export TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:31:40.3691718Z + TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:31:40.3691957Z + export VALGRIND=ON 2025-12-04T09:31:40.3692169Z + VALGRIND=ON 2025-12-04T09:31:40.3692392Z + [[ linux-jammy-rocm-py3.10 == *clang9* ]] 2025-12-04T09:31:40.3692680Z + [[ linux-jammy-rocm-py3.10 == *xpu* ]] 2025-12-04T09:31:40.3692935Z + detect_cuda_arch 2025-12-04T09:31:40.3693154Z + [[ linux-jammy-rocm-py3.10 == *cuda* ]] 2025-12-04T09:31:40.3693434Z + [[ linux-jammy-rocm-py3.10 == *s390x* ]] 2025-12-04T09:31:40.3693679Z + [[ 1 == \1 ]] 2025-12-04T09:31:40.3693876Z + ulimit -c 0 2025-12-04T09:31:40.3694088Z + [[ linux-jammy-rocm-py3.10 != *bazel* ]] 2025-12-04T09:31:40.3694371Z ++ realpath build/custom_test_artifacts 2025-12-04T09:31:40.3698822Z + CUSTOM_TEST_ARTIFACT_BUILD_DIR=/var/lib/jenkins/pytorch/build/custom_test_artifacts 2025-12-04T09:31:40.3699231Z + [[ -n '' ]] 2025-12-04T09:31:40.3699454Z + echo 'Environment variables' 2025-12-04T09:31:40.3699709Z Environment variables 2025-12-04T09:31:40.3701658Z + env 2025-12-04T09:31:40.3705039Z GITHUB_WORKSPACE=/home/runner/_work/pytorch/pytorch 2025-12-04T09:31:40.3705309Z CONTINUE_THROUGH_ERROR=True 2025-12-04T09:31:40.3705527Z BUILD_ENVIRONMENT=linux-jammy-rocm-py3.10 2025-12-04T09:31:40.3705807Z HOSTNAME=linux.rocm.gpu.gfx942.1.b-gwk9b-runner-68hm6 2025-12-04T09:31:40.3706210Z GITHUB_PATH=/home/runner/_work/_temp/_runner_file_commands/add_path_6a3264ba-5c74-4e58-a72b-bbfb124cda02 2025-12-04T09:31:40.3706545Z GITHUB_ACTION=__run_2 2025-12-04T09:31:40.3706732Z PYTORCH_TEST_CUDA_MEM_LEAK_CHECK=0 2025-12-04T09:31:40.3706992Z GITHUB_RUN_NUMBER=689 2025-12-04T09:31:40.3707158Z TEST_CONFIG=default 2025-12-04T09:31:40.3707372Z RUNNER_NAME=linux.rocm.gpu.gfx942.1.b-gwk9b-runner-68hm6 2025-12-04T09:31:40.3707632Z GITHUB_REPOSITORY_OWNER_ID=21003710 2025-12-04T09:31:40.3707841Z AWS_DEFAULT_REGION=us-east-1 2025-12-04T09:31:40.3708085Z RUNNER_ARTIFACT_DIR=/home/runner/_work/_temp/artifacts 2025-12-04T09:31:40.3708332Z GITHUB_TRIGGERING_ACTOR=pytorchmergebot 2025-12-04T09:31:40.3708543Z GITHUB_REF_TYPE=branch 2025-12-04T09:31:40.3708740Z BASE_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:31:40.3709154Z HUGGING_FACE_HUB_TOKEN=*** 2025-12-04T09:31:40.3709406Z *** 2025-12-04T09:31:40.3709557Z GITHUB_REPOSITORY_ID=65600975 2025-12-04T09:31:40.3709736Z GITHUB_ACTIONS=true 2025-12-04T09:31:40.3709925Z SHA1=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:31:40.3710189Z GITHUB_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:31:40.3710550Z GITHUB_WORKFLOW_REF=pytorch/pytorch/.github/workflows/trunk-rocm-mi300.yml@refs/heads/main 2025-12-04T09:31:40.3710864Z UCC_HOME=/usr 2025-12-04T09:31:40.3711021Z TORCH_SERIALIZATION_DEBUG=1 2025-12-04T09:31:40.3711228Z RUNNER_ENVIRONMENT=self-hosted 2025-12-04T09:31:40.3711404Z VERBOSE_TEST_LOGS=False 2025-12-04T09:31:40.3711542Z GITHUB_REF=refs/heads/main 2025-12-04T09:31:40.3711685Z RUNNER_OS=Linux 2025-12-04T09:31:40.3711816Z SHARD_NUMBER=4 2025-12-04T09:31:40.3712144Z GITHUB_REF_PROTECTED=true 2025-12-04T09:31:40.3712298Z RUNNER_MANUALLY_TRAP_SIG=1 2025-12-04T09:31:40.3712445Z HOME=/var/lib/jenkins 2025-12-04T09:31:40.3712604Z GITHUB_API_URL=https://api.github.com 2025-12-04T09:31:40.3712783Z PYTORCH_TEST_RERUN_DISABLED_TESTS=1 2025-12-04T09:31:40.3712954Z RUNNER_DOCS_DIR=/home/runner/_work/_temp/docs 2025-12-04T09:31:40.3713114Z LANG=C.UTF-8 2025-12-04T09:31:40.3713263Z UCX_COMMIT=29831d319e6be55cb8c768ca61de335c934ca39e 2025-12-04T09:31:40.3713457Z PYTORCH_TEST_WITH_ROCM=1 2025-12-04T09:31:40.3713634Z RUNNER_TRACKING_ID=github_c5d5cb89-1e45-4146-be76-5d780902345a 2025-12-04T09:31:40.3713828Z RUNNER_ARCH=X64 2025-12-04T09:31:40.3713967Z RUNNER_TEMP=/home/runner/_work/_temp 2025-12-04T09:31:40.3714121Z NUM_TEST_SHARDS=6 2025-12-04T09:31:40.3714238Z UCX_HOME=/usr 2025-12-04T09:31:40.3714488Z GITHUB_STATE=/home/runner/_work/_temp/_runner_file_commands/save_state_6a3264ba-5c74-4e58-a72b-bbfb124cda02 2025-12-04T09:31:40.3714904Z JOB_NAME=linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable) 2025-12-04T09:31:40.3715193Z MAGMA_HOME=/opt/rocm/magma 2025-12-04T09:31:40.3715437Z GITHUB_ENV=/home/runner/_work/_temp/_runner_file_commands/set_env_6a3264ba-5c74-4e58-a72b-bbfb124cda02 2025-12-04T09:31:40.3715761Z GITHUB_EVENT_PATH=/home/runner/_work/_temp/_github_workflow/event.json 2025-12-04T09:31:40.3715970Z GITHUB_EVENT_NAME=schedule 2025-12-04T09:31:40.3716174Z GITHUB_ACTIONS_RUNNER_EXTRA_USER_AGENT=actions-runner-controller/0.12.1 2025-12-04T09:31:40.3716383Z DASHBOARD_TAG= 2025-12-04T09:31:40.3716507Z GITHUB_RUN_ID=19922849170 2025-12-04T09:31:40.3716845Z GITHUB_STEP_SUMMARY=/home/runner/_work/_temp/_runner_file_commands/step_summary_6a3264ba-5c74-4e58-a72b-bbfb124cda02 2025-12-04T09:31:40.3717143Z GITHUB_ACTOR=pytorchmergebot 2025-12-04T09:31:40.3717282Z PR_NUMBER= 2025-12-04T09:31:40.3717404Z GITHUB_RUN_ATTEMPT=1 2025-12-04T09:31:40.3717528Z VALGRIND=ON 2025-12-04T09:31:40.3717665Z ANACONDA_PYTHON_VERSION=3.10 2025-12-04T09:31:40.3717843Z GITHUB_GRAPHQL_URL=https://api.github.com/graphql 2025-12-04T09:31:40.3718085Z TERM=vt100 2025-12-04T09:31:40.3718207Z INSTALLED_VISION=yes 2025-12-04T09:31:40.3718335Z BRANCH=main 2025-12-04T09:31:40.3718459Z OPENSSL_ROOT_DIR=/opt/openssl 2025-12-04T09:31:40.3718603Z TESTS_TO_INCLUDE= 2025-12-04T09:31:40.3718816Z GITHUB_ACTION_PATH=/home/runner/_work/pytorch/pytorch/./.github/actions/setup-rocm 2025-12-04T09:31:40.3719064Z GITHUB_SERVER_URL=https://github.com 2025-12-04T09:31:40.3719249Z PYTORCH_ROCM_ARCH=gfx90a;gfx942;gfx950;gfx1100 2025-12-04T09:31:40.3719446Z UCC_COMMIT=9f4b242cbbd8b1462cbc732eb29316cdfa124b77 2025-12-04T09:31:40.3719614Z REENABLED_ISSUES= 2025-12-04T09:31:40.3719727Z SHLVL=1 2025-12-04T09:31:40.3719841Z MAX_JOBS=126 2025-12-04T09:31:40.3720029Z RUNNER_TEST_RESULTS_DIR=/home/runner/_work/_temp/test-results 2025-12-04T09:31:40.3720223Z GITHUB_ACTOR_ID=97764156 2025-12-04T09:31:40.3720374Z RUNNER_TOOL_CACHE=/home/runner/_work/_tool 2025-12-04T09:31:40.3720585Z GITHUB_WORKFLOW_SHA=ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T09:31:40.3720786Z GITHUB_REF_NAME=main 2025-12-04T09:31:40.3720914Z ROCM_PATH=/opt/rocm 2025-12-04T09:31:40.3721042Z GITHUB_JOB=test 2025-12-04T09:31:40.3721161Z NO_TEST_TIMEOUT=False 2025-12-04T09:31:40.3721354Z GITHUB_REPOSITORY=pytorch/pytorch 2025-12-04T09:31:40.3721474Z LC_ALL=C.UTF-8 2025-12-04T09:31:40.3721575Z GITHUB_RETENTION_DAYS=90 2025-12-04T09:31:40.3721699Z RUNNER_WORKSPACE=/home/runner/_work/pytorch 2025-12-04T09:31:40.3721834Z OPENSSL_DIR=/opt/openssl 2025-12-04T09:31:40.3721951Z GITHUB_ACTION_REPOSITORY= 2025-12-04T09:31:40.3722329Z PATH=/opt/cache/bin:/opt/rocm/llvm/bin:/opt/rocm/opencl/bin:/opt/rocm/hip/bin:/opt/rocm/hcc/bin:/opt/rocm/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:31:40.3722694Z GITHUB_BASE_REF= 2025-12-04T09:31:40.3722794Z CI=true 2025-12-04T09:31:40.3722894Z GITHUB_REPOSITORY_OWNER=pytorch 2025-12-04T09:31:40.3723070Z JOB_ID=57116213149 2025-12-04T09:31:40.3723170Z GITHUB_HEAD_REF= 2025-12-04T09:31:40.3723276Z GITHUB_ACTION_REF= 2025-12-04T09:31:40.3723377Z TEST_SHOWLOCALS=False 2025-12-04T09:31:40.3723492Z GITHUB_WORKFLOW=trunk-rocm-mi300 2025-12-04T09:31:40.3723618Z DEBIAN_FRONTEND=noninteractive 2025-12-04T09:31:40.3723834Z GITHUB_OUTPUT=/home/runner/_work/_temp/_runner_file_commands/set_output_6a3264ba-5c74-4e58-a72b-bbfb124cda02 2025-12-04T09:31:40.3724057Z NO_TD=False 2025-12-04T09:31:40.3724152Z OLDPWD=/var/lib/jenkins 2025-12-04T09:31:40.3724259Z _=/usr/bin/env 2025-12-04T09:31:40.3724360Z + echo 'Testing pytorch' 2025-12-04T09:31:40.3724469Z Testing pytorch 2025-12-04T09:31:40.3724572Z + export LANG=C.UTF-8 2025-12-04T09:31:40.3724675Z + LANG=C.UTF-8 2025-12-04T09:31:40.3724769Z + PR_NUMBER= 2025-12-04T09:31:40.3724871Z + [[ default == \d\e\f\a\u\l\t ]] 2025-12-04T09:31:40.3724994Z + export CUDA_VISIBLE_DEVICES=0 2025-12-04T09:31:40.3725115Z + CUDA_VISIBLE_DEVICES=0 2025-12-04T09:31:40.3725237Z + export HIP_VISIBLE_DEVICES=0 2025-12-04T09:31:40.3725357Z + HIP_VISIBLE_DEVICES=0 2025-12-04T09:31:40.3725475Z + [[ default == \d\i\s\t\r\i\b\u\t\e\d ]] 2025-12-04T09:31:40.3725604Z + [[ default == \s\l\o\w ]] 2025-12-04T09:31:40.3725742Z + [[ linux-jammy-rocm-py3.10 == *slow-gradcheck* ]] 2025-12-04T09:31:40.3725895Z + [[ linux-jammy-rocm-py3.10 == *cuda* ]] 2025-12-04T09:31:40.3726032Z + [[ linux-jammy-rocm-py3.10 == *rocm* ]] 2025-12-04T09:31:40.3726173Z + export PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-12-04T09:31:40.3726320Z + PYTORCH_TESTING_DEVICE_ONLY_FOR=cuda 2025-12-04T09:31:40.3726476Z + [[ default == *crossref* ]] 2025-12-04T09:31:40.3726600Z + [[ linux-jammy-rocm-py3.10 == *rocm* ]] 2025-12-04T09:31:40.3726730Z + export VALGRIND=OFF 2025-12-04T09:31:40.3726867Z + VALGRIND=OFF 2025-12-04T09:31:40.3726960Z + rocminfo 2025-12-04T09:31:40.3832453Z ROCk module version 6.12.12 is loaded 2025-12-04T09:31:40.4204802Z ===================== 2025-12-04T09:31:40.4205175Z HSA System Attributes 2025-12-04T09:31:40.4205497Z ===================== 2025-12-04T09:31:40.4207085Z Runtime Version: 1.18 2025-12-04T09:31:40.4207404Z Runtime Ext Version: 1.14 2025-12-04T09:31:40.4207741Z System Timestamp Freq.: 1000.000000MHz 2025-12-04T09:31:40.4208258Z Sig. Max Wait Duration: 18446744073709551615 (0xFFFFFFFFFFFFFFFF) (timestamp count) 2025-12-04T09:31:40.4208823Z Machine Model: LARGE 2025-12-04T09:31:40.4209276Z System Endianness: LITTLE 2025-12-04T09:31:40.4209670Z Mwaitx: DISABLED 2025-12-04T09:31:40.4209994Z XNACK enabled: NO 2025-12-04T09:31:40.4210300Z DMAbuf Support: YES 2025-12-04T09:31:40.4210601Z VMM Support: YES 2025-12-04T09:31:40.4210793Z 2025-12-04T09:31:40.4210903Z ========== 2025-12-04T09:31:40.4211171Z HSA Agents 2025-12-04T09:31:40.4211298Z ========== 2025-12-04T09:31:40.4211402Z ******* 2025-12-04T09:31:40.4211508Z Agent 1 2025-12-04T09:31:40.4211624Z ******* 2025-12-04T09:31:40.4211756Z Name: AMD EPYC 9575F 64-Core Processor 2025-12-04T09:31:40.4211917Z Uuid: CPU-XX 2025-12-04T09:31:40.4212088Z Marketing Name: AMD EPYC 9575F 64-Core Processor 2025-12-04T09:31:40.4212260Z Vendor Name: CPU 2025-12-04T09:31:40.4212426Z Feature: None specified 2025-12-04T09:31:40.4212589Z Profile: FULL_PROFILE 2025-12-04T09:31:40.4212758Z Float Round Mode: NEAR 2025-12-04T09:31:40.4212927Z Max Queue Number: 0(0x0) 2025-12-04T09:31:40.4213092Z Queue Min Size: 0(0x0) 2025-12-04T09:31:40.4213252Z Queue Max Size: 0(0x0) 2025-12-04T09:31:40.4213494Z Queue Type: MULTI 2025-12-04T09:31:40.4213662Z Node: 0 2025-12-04T09:31:40.4213824Z Device Type: CPU 2025-12-04T09:31:40.4213975Z Cache Info: 2025-12-04T09:31:40.4214107Z L1: 49152(0xc000) KB 2025-12-04T09:31:40.4214259Z Chip ID: 0(0x0) 2025-12-04T09:31:40.4214417Z ASIC Revision: 0(0x0) 2025-12-04T09:31:40.4214583Z Cacheline Size: 64(0x40) 2025-12-04T09:31:40.4214754Z Max Clock Freq. (MHz): 3300 2025-12-04T09:31:40.4214912Z BDFID: 0 2025-12-04T09:31:40.4215070Z Internal Node ID: 0 2025-12-04T09:31:40.4215238Z Compute Unit: 64 2025-12-04T09:31:40.4215403Z SIMDs per CU: 0 2025-12-04T09:31:40.4215569Z Shader Engines: 0 2025-12-04T09:31:40.4215737Z Shader Arrs. per Eng.: 0 2025-12-04T09:31:40.4215911Z WatchPts on Addr. Ranges:1 2025-12-04T09:31:40.4216071Z Memory Properties: 2025-12-04T09:31:40.4216200Z Features: None 2025-12-04T09:31:40.4216323Z Pool Info: 2025-12-04T09:31:40.4216446Z Pool 1 2025-12-04T09:31:40.4216597Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-12-04T09:31:40.4216837Z Size: 1584734448(0x5e7520f0) KB 2025-12-04T09:31:40.4217010Z Allocatable: TRUE 2025-12-04T09:31:40.4217185Z Alloc Granule: 4KB 2025-12-04T09:31:40.4217379Z Alloc Recommended Granule:4KB 2025-12-04T09:31:40.4217605Z Alloc Alignment: 4KB 2025-12-04T09:31:40.4217788Z Accessible by all: TRUE 2025-12-04T09:31:40.4217945Z Pool 2 2025-12-04T09:31:40.4218096Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-12-04T09:31:40.4218264Z Size: 1584734448(0x5e7520f0) KB 2025-12-04T09:31:40.4218429Z Allocatable: TRUE 2025-12-04T09:31:40.4218607Z Alloc Granule: 4KB 2025-12-04T09:31:40.4218788Z Alloc Recommended Granule:4KB 2025-12-04T09:31:40.4218967Z Alloc Alignment: 4KB 2025-12-04T09:31:40.4219144Z Accessible by all: TRUE 2025-12-04T09:31:40.4219302Z Pool 3 2025-12-04T09:31:40.4219452Z Segment: GLOBAL; FLAGS: KERNARG, FINE GRAINED 2025-12-04T09:31:40.4219621Z Size: 1584734448(0x5e7520f0) KB 2025-12-04T09:31:40.4219870Z Allocatable: TRUE 2025-12-04T09:31:40.4220048Z Alloc Granule: 4KB 2025-12-04T09:31:40.4220233Z Alloc Recommended Granule:4KB 2025-12-04T09:31:40.4220415Z Alloc Alignment: 4KB 2025-12-04T09:31:40.4220593Z Accessible by all: TRUE 2025-12-04T09:31:40.4220746Z Pool 4 2025-12-04T09:31:40.4220889Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-12-04T09:31:40.4221055Z Size: 1584734448(0x5e7520f0) KB 2025-12-04T09:31:40.4221278Z Allocatable: TRUE 2025-12-04T09:31:40.4221445Z Alloc Granule: 4KB 2025-12-04T09:31:40.4221617Z Alloc Recommended Granule:4KB 2025-12-04T09:31:40.4221786Z Alloc Alignment: 4KB 2025-12-04T09:31:40.4221949Z Accessible by all: TRUE 2025-12-04T09:31:40.4222092Z ISA Info: 2025-12-04T09:31:40.4222200Z ******* 2025-12-04T09:31:40.4222302Z Agent 2 2025-12-04T09:31:40.4222409Z ******* 2025-12-04T09:31:40.4222531Z Name: AMD EPYC 9575F 64-Core Processor 2025-12-04T09:31:40.4222682Z Uuid: CPU-XX 2025-12-04T09:31:40.4222840Z Marketing Name: AMD EPYC 9575F 64-Core Processor 2025-12-04T09:31:40.4223004Z Vendor Name: CPU 2025-12-04T09:31:40.4223165Z Feature: None specified 2025-12-04T09:31:40.4223323Z Profile: FULL_PROFILE 2025-12-04T09:31:40.4223481Z Float Round Mode: NEAR 2025-12-04T09:31:40.4223641Z Max Queue Number: 0(0x0) 2025-12-04T09:31:40.4223797Z Queue Min Size: 0(0x0) 2025-12-04T09:31:40.4223951Z Queue Max Size: 0(0x0) 2025-12-04T09:31:40.4224106Z Queue Type: MULTI 2025-12-04T09:31:40.4224254Z Node: 1 2025-12-04T09:31:40.4224402Z Device Type: CPU 2025-12-04T09:31:40.4224540Z Cache Info: 2025-12-04T09:31:40.4224662Z L1: 49152(0xc000) KB 2025-12-04T09:31:40.4224809Z Chip ID: 0(0x0) 2025-12-04T09:31:40.4224989Z ASIC Revision: 0(0x0) 2025-12-04T09:31:40.4225150Z Cacheline Size: 64(0x40) 2025-12-04T09:31:40.4225312Z Max Clock Freq. (MHz): 3300 2025-12-04T09:31:40.4225461Z BDFID: 0 2025-12-04T09:31:40.4225613Z Internal Node ID: 1 2025-12-04T09:31:40.4225769Z Compute Unit: 64 2025-12-04T09:31:40.4225923Z SIMDs per CU: 0 2025-12-04T09:31:40.4226077Z Shader Engines: 0 2025-12-04T09:31:40.4226236Z Shader Arrs. per Eng.: 0 2025-12-04T09:31:40.4226400Z WatchPts on Addr. Ranges:1 2025-12-04T09:31:40.4226551Z Memory Properties: 2025-12-04T09:31:40.4226670Z Features: None 2025-12-04T09:31:40.4226826Z Pool Info: 2025-12-04T09:31:40.4226937Z Pool 1 2025-12-04T09:31:40.4227069Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-12-04T09:31:40.4227225Z Size: 1585355632(0x5e7e9b70) KB 2025-12-04T09:31:40.4227381Z Allocatable: TRUE 2025-12-04T09:31:40.4227543Z Alloc Granule: 4KB 2025-12-04T09:31:40.4227713Z Alloc Recommended Granule:4KB 2025-12-04T09:31:40.4227879Z Alloc Alignment: 4KB 2025-12-04T09:31:40.4228046Z Accessible by all: TRUE 2025-12-04T09:31:40.4228188Z Pool 2 2025-12-04T09:31:40.4228323Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-12-04T09:31:40.4228519Z Size: 1585355632(0x5e7e9b70) KB 2025-12-04T09:31:40.4228677Z Allocatable: TRUE 2025-12-04T09:31:40.4228840Z Alloc Granule: 4KB 2025-12-04T09:31:40.4229009Z Alloc Recommended Granule:4KB 2025-12-04T09:31:40.4229178Z Alloc Alignment: 4KB 2025-12-04T09:31:40.4229342Z Accessible by all: TRUE 2025-12-04T09:31:40.4229486Z Pool 3 2025-12-04T09:31:40.4229618Z Segment: GLOBAL; FLAGS: KERNARG, FINE GRAINED 2025-12-04T09:31:40.4229773Z Size: 1585355632(0x5e7e9b70) KB 2025-12-04T09:31:40.4229927Z Allocatable: TRUE 2025-12-04T09:31:40.4230087Z Alloc Granule: 4KB 2025-12-04T09:31:40.4230258Z Alloc Recommended Granule:4KB 2025-12-04T09:31:40.4230427Z Alloc Alignment: 4KB 2025-12-04T09:31:40.4230590Z Accessible by all: TRUE 2025-12-04T09:31:40.4230732Z Pool 4 2025-12-04T09:31:40.4230867Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-12-04T09:31:40.4231019Z Size: 1585355632(0x5e7e9b70) KB 2025-12-04T09:31:40.4231171Z Allocatable: TRUE 2025-12-04T09:31:40.4231332Z Alloc Granule: 4KB 2025-12-04T09:31:40.4231498Z Alloc Recommended Granule:4KB 2025-12-04T09:31:40.4231665Z Alloc Alignment: 4KB 2025-12-04T09:31:40.4231830Z Accessible by all: TRUE 2025-12-04T09:31:40.4231981Z ISA Info: 2025-12-04T09:31:40.4232125Z ******* 2025-12-04T09:31:40.4232230Z Agent 3 2025-12-04T09:31:40.4232335Z ******* 2025-12-04T09:31:40.4232450Z Name: gfx942 2025-12-04T09:31:40.4232600Z Uuid: GPU-8052c781c18bf05c 2025-12-04T09:31:40.4232754Z Marketing Name: 2025-12-04T09:31:40.4232914Z Vendor Name: AMD 2025-12-04T09:31:40.4233071Z Feature: KERNEL_DISPATCH 2025-12-04T09:31:40.4233228Z Profile: BASE_PROFILE 2025-12-04T09:31:40.4233389Z Float Round Mode: NEAR 2025-12-04T09:31:40.4233548Z Max Queue Number: 128(0x80) 2025-12-04T09:31:40.4233708Z Queue Min Size: 64(0x40) 2025-12-04T09:31:40.4233863Z Queue Max Size: 131072(0x20000) 2025-12-04T09:31:40.4234025Z Queue Type: MULTI 2025-12-04T09:31:40.4234172Z Node: 2 2025-12-04T09:31:40.4234320Z Device Type: GPU 2025-12-04T09:31:40.4234458Z Cache Info: 2025-12-04T09:31:40.4244066Z L1: 32(0x20) KB 2025-12-04T09:31:40.4244254Z L2: 4096(0x1000) KB 2025-12-04T09:31:40.4244399Z L3: 262144(0x40000) KB 2025-12-04T09:31:40.4244549Z Chip ID: 29861(0x74a5) 2025-12-04T09:31:40.4244714Z ASIC Revision: 1(0x1) 2025-12-04T09:31:40.4244884Z Cacheline Size: 128(0x80) 2025-12-04T09:31:40.4245122Z Max Clock Freq. (MHz): 2100 2025-12-04T09:31:40.4245290Z BDFID: 5376 2025-12-04T09:31:40.4245449Z Internal Node ID: 2 2025-12-04T09:31:40.4245613Z Compute Unit: 304 2025-12-04T09:31:40.4245774Z SIMDs per CU: 4 2025-12-04T09:31:40.4245936Z Shader Engines: 32 2025-12-04T09:31:40.4246100Z Shader Arrs. per Eng.: 1 2025-12-04T09:31:40.4246269Z WatchPts on Addr. Ranges:4 2025-12-04T09:31:40.4246440Z Coherent Host Access: FALSE 2025-12-04T09:31:40.4246594Z Memory Properties: 2025-12-04T09:31:40.4246725Z Features: KERNEL_DISPATCH 2025-12-04T09:31:40.4246931Z Fast F16 Operation: TRUE 2025-12-04T09:31:40.4247104Z Wavefront Size: 64(0x40) 2025-12-04T09:31:40.4247277Z Workgroup Max Size: 1024(0x400) 2025-12-04T09:31:40.4247431Z Workgroup Max Size per Dimension: 2025-12-04T09:31:40.4247567Z x 1024(0x400) 2025-12-04T09:31:40.4247709Z y 1024(0x400) 2025-12-04T09:31:40.4247851Z z 1024(0x400) 2025-12-04T09:31:40.4248006Z Max Waves Per CU: 32(0x20) 2025-12-04T09:31:40.4248174Z Max Work-item Per CU: 2048(0x800) 2025-12-04T09:31:40.4248340Z Grid Max Size: 4294967295(0xffffffff) 2025-12-04T09:31:40.4248492Z Grid Max Size per Dimension: 2025-12-04T09:31:40.4248621Z x 2147483647(0x7fffffff) 2025-12-04T09:31:40.4248768Z y 65535(0xffff) 2025-12-04T09:31:40.4248907Z z 65535(0xffff) 2025-12-04T09:31:40.4249113Z Max fbarriers/Workgrp: 32 2025-12-04T09:31:40.4249340Z Packet Processor uCode:: 185 2025-12-04T09:31:40.4249513Z SDMA engine uCode:: 24 2025-12-04T09:31:40.4249682Z IOMMU Support:: None 2025-12-04T09:31:40.4249831Z Pool Info: 2025-12-04T09:31:40.4249950Z Pool 1 2025-12-04T09:31:40.4250094Z Segment: GLOBAL; FLAGS: COARSE GRAINED 2025-12-04T09:31:40.4250261Z Size: 268419072(0xfffc000) KB 2025-12-04T09:31:40.4250424Z Allocatable: TRUE 2025-12-04T09:31:40.4250593Z Alloc Granule: 4KB 2025-12-04T09:31:40.4250773Z Alloc Recommended Granule:2048KB 2025-12-04T09:31:40.4250953Z Alloc Alignment: 4KB 2025-12-04T09:31:40.4251126Z Accessible by all: FALSE 2025-12-04T09:31:40.4251275Z Pool 2 2025-12-04T09:31:40.4251417Z Segment: GLOBAL; FLAGS: EXTENDED FINE GRAINED 2025-12-04T09:31:40.4251577Z Size: 268419072(0xfffc000) KB 2025-12-04T09:31:40.4251737Z Allocatable: TRUE 2025-12-04T09:31:40.4251905Z Alloc Granule: 4KB 2025-12-04T09:31:40.4252077Z Alloc Recommended Granule:2048KB 2025-12-04T09:31:40.4252252Z Alloc Alignment: 4KB 2025-12-04T09:31:40.4252421Z Accessible by all: FALSE 2025-12-04T09:31:40.4252605Z Pool 3 2025-12-04T09:31:40.4252745Z Segment: GLOBAL; FLAGS: FINE GRAINED 2025-12-04T09:31:40.4252908Z Size: 268419072(0xfffc000) KB 2025-12-04T09:31:40.4253068Z Allocatable: TRUE 2025-12-04T09:31:40.4253235Z Alloc Granule: 4KB 2025-12-04T09:31:40.4253406Z Alloc Recommended Granule:2048KB 2025-12-04T09:31:40.4253579Z Alloc Alignment: 4KB 2025-12-04T09:31:40.4253748Z Accessible by all: FALSE 2025-12-04T09:31:40.4253891Z Pool 4 2025-12-04T09:31:40.4254026Z Segment: GROUP 2025-12-04T09:31:40.4254180Z Size: 64(0x40) KB 2025-12-04T09:31:40.4254343Z Allocatable: FALSE 2025-12-04T09:31:40.4254509Z Alloc Granule: 0KB 2025-12-04T09:31:40.4254684Z Alloc Recommended Granule:0KB 2025-12-04T09:31:40.4254857Z Alloc Alignment: 0KB 2025-12-04T09:31:40.4255026Z Accessible by all: FALSE 2025-12-04T09:31:40.4255178Z ISA Info: 2025-12-04T09:31:40.4255294Z ISA 1 2025-12-04T09:31:40.4255441Z Name: amdgcn-amd-amdhsa--gfx942:sramecc+:xnack- 2025-12-04T09:31:40.4255618Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-12-04T09:31:40.4255791Z Profiles: HSA_PROFILE_BASE 2025-12-04T09:31:40.4255966Z Default Rounding Mode: NEAR 2025-12-04T09:31:40.4256141Z Default Rounding Mode: NEAR 2025-12-04T09:31:40.4256311Z Fast f16: TRUE 2025-12-04T09:31:40.4256508Z Workgroup Max Size: 1024(0x400) 2025-12-04T09:31:40.4256665Z Workgroup Max Size per Dimension: 2025-12-04T09:31:40.4256872Z x 1024(0x400) 2025-12-04T09:31:40.4257016Z y 1024(0x400) 2025-12-04T09:31:40.4257155Z z 1024(0x400) 2025-12-04T09:31:40.4257307Z Grid Max Size: 4294967295(0xffffffff) 2025-12-04T09:31:40.4257459Z Grid Max Size per Dimension: 2025-12-04T09:31:40.4257593Z x 2147483647(0x7fffffff) 2025-12-04T09:31:40.4257734Z y 65535(0xffff) 2025-12-04T09:31:40.4257873Z z 65535(0xffff) 2025-12-04T09:31:40.4258034Z FBarrier Max Size: 32 2025-12-04T09:31:40.4258188Z ISA 2 2025-12-04T09:31:40.4258344Z Name: amdgcn-amd-amdhsa--gfx9-4-generic:sramecc+:xnack- 2025-12-04T09:31:40.4258535Z Machine Models: HSA_MACHINE_MODEL_LARGE 2025-12-04T09:31:40.4258711Z Profiles: HSA_PROFILE_BASE 2025-12-04T09:31:40.4258886Z Default Rounding Mode: NEAR 2025-12-04T09:31:40.4259064Z Default Rounding Mode: NEAR 2025-12-04T09:31:40.4259231Z Fast f16: TRUE 2025-12-04T09:31:40.4259397Z Workgroup Max Size: 1024(0x400) 2025-12-04T09:31:40.4259554Z Workgroup Max Size per Dimension: 2025-12-04T09:31:40.4259697Z x 1024(0x400) 2025-12-04T09:31:40.4259871Z y 1024(0x400) 2025-12-04T09:31:40.4260014Z z 1024(0x400) 2025-12-04T09:31:40.4260167Z Grid Max Size: 4294967295(0xffffffff) 2025-12-04T09:31:40.4260318Z Grid Max Size per Dimension: 2025-12-04T09:31:40.4260453Z x 2147483647(0x7fffffff) 2025-12-04T09:31:40.4260598Z y 65535(0xffff) 2025-12-04T09:31:40.4260740Z z 65535(0xffff) 2025-12-04T09:31:40.4260898Z FBarrier Max Size: 32 2025-12-04T09:31:40.4261045Z *** Done *** 2025-12-04T09:31:40.4278620Z + rocminfo 2025-12-04T09:31:40.4281648Z + grep -E 'Name:.*\sgfx|Marketing' 2025-12-04T09:31:40.4770043Z Marketing Name: AMD EPYC 9575F 64-Core Processor 2025-12-04T09:31:40.4770493Z Marketing Name: AMD EPYC 9575F 64-Core Processor 2025-12-04T09:31:40.4770890Z Name: gfx942 2025-12-04T09:31:40.4771230Z Marketing Name: 2025-12-04T09:31:40.4813313Z + MAYBE_ROCM=rocm/ 2025-12-04T09:31:40.4813548Z + [[ linux-jammy-rocm-py3.10 == *xpu* ]] 2025-12-04T09:31:40.4813823Z + [[ linux-jammy-rocm-py3.10 != *-bazel-* ]] 2025-12-04T09:31:40.4814069Z + pip_install ninja==1.10.2 2025-12-04T09:31:40.4815305Z + pip_install_pkg='python3 -m pip install --progress-bar off' 2025-12-04T09:31:40.4815643Z + python3 -m pip install --progress-bar off ninja==1.10.2 2025-12-04T09:31:40.6756844Z Collecting ninja==1.10.2 2025-12-04T09:31:40.7016583Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl.metadata (5.0 kB) 2025-12-04T09:31:40.7098941Z Downloading ninja-1.10.2-py2.py3-none-manylinux_2_5_x86_64.manylinux1_x86_64.whl (108 kB) 2025-12-04T09:31:40.8789854Z Installing collected packages: ninja 2025-12-04T09:31:40.8790145Z Attempting uninstall: ninja 2025-12-04T09:31:40.8792468Z Found existing installation: ninja 1.11.1.4 2025-12-04T09:31:40.8804930Z Uninstalling ninja-1.11.1.4: 2025-12-04T09:31:40.8901704Z Successfully uninstalled ninja-1.11.1.4 2025-12-04T09:31:40.9023881Z Successfully installed ninja-1.10.2 2025-12-04T09:31:40.9516071Z + export PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/opt/rocm/llvm/bin:/opt/rocm/opencl/bin:/opt/rocm/hip/bin:/opt/rocm/hcc/bin:/opt/rocm/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:31:40.9518138Z + PATH=/var/lib/jenkins/.local/bin:/opt/cache/bin:/opt/rocm/llvm/bin:/opt/rocm/opencl/bin:/opt/rocm/hip/bin:/opt/rocm/hcc/bin:/opt/rocm/bin:/opt/conda/envs/py_3.10/bin:/opt/conda/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin 2025-12-04T09:31:40.9519274Z + [[ linux-jammy-rocm-py3.10 == *aarch64* ]] 2025-12-04T09:31:40.9519675Z + [[ linux-jammy-rocm-py3.10 == *asan* ]] 2025-12-04T09:31:40.9520117Z + [[ linux-jammy-rocm-py3.10 == *-debug* ]] 2025-12-04T09:31:40.9520514Z + [[ linux-jammy-rocm-py3.10 != *-bazel-* ]] 2025-12-04T09:31:40.9521047Z + echo 'We are not in debug mode: linux-jammy-rocm-py3.10. Expect the assertion to pass' 2025-12-04T09:31:40.9521342Z We are not in debug mode: linux-jammy-rocm-py3.10. Expect the assertion to pass 2025-12-04T09:31:40.9521515Z + cd test 2025-12-04T09:31:40.9521670Z + python -c 'import torch; torch._C._crash_if_debug_asserts_fail(424242)' 2025-12-04T09:31:42.6542120Z + [[ default == \n\o\g\p\u\_\N\O\_\A\V\X\2 ]] 2025-12-04T09:31:42.6542428Z + [[ default == \n\o\g\p\u\_\A\V\X\5\1\2 ]] 2025-12-04T09:31:42.6542671Z + [[ default == \l\e\g\a\c\y\_\n\v\i\d\i\a\_\d\r\i\v\e\r ]] 2025-12-04T09:31:42.6545322Z + DYNAMO_BENCHMARK_FLAGS=() 2025-12-04T09:31:42.6546884Z + [[ default == *pr_time_benchmarks* ]] 2025-12-04T09:31:42.6547127Z + [[ default == *dynamo_eager* ]] 2025-12-04T09:31:42.6547325Z + [[ default == *aot_eager* ]] 2025-12-04T09:31:42.6548815Z + [[ default == *aot_inductor* ]] 2025-12-04T09:31:42.6549018Z + [[ default == *max_autotune_inductor* ]] 2025-12-04T09:31:42.6549249Z + [[ default == *inductor* ]] 2025-12-04T09:31:42.6549433Z + [[ default == *dynamic* ]] 2025-12-04T09:31:42.6549618Z + [[ default == *cpu* ]] 2025-12-04T09:31:42.6549788Z + [[ default == *xpu* ]] 2025-12-04T09:31:42.6549996Z + DYNAMO_BENCHMARK_FLAGS+=(--device cuda) 2025-12-04T09:31:42.6558144Z + [[ linux-jammy-rocm-py3.10 == *libtorch* ]] 2025-12-04T09:31:42.6558352Z + [[ linux-jammy-rocm-py3.10 == *-bazel-* ]] 2025-12-04T09:31:42.6563545Z + cd test 2025-12-04T09:31:42.6563835Z + python -c 'import torch; print(torch.__config__.show())' 2025-12-04T09:31:44.2659185Z PyTorch built with: 2025-12-04T09:31:44.2659569Z - GCC 11.4 2025-12-04T09:31:44.2659844Z - C++ Version: 201703 2025-12-04T09:31:44.2660474Z - Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:31:44.2661149Z - Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:31:44.2661337Z - OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:31:44.2661477Z - LAPACK is enabled (usually provided by MKL) 2025-12-04T09:31:44.2661608Z - NNPACK is enabled 2025-12-04T09:31:44.2661721Z - CPU capability usage: AVX512 2025-12-04T09:31:44.2661840Z - HIP Runtime 7.1.25424 2025-12-04T09:31:44.2661948Z - MIOpen 3.5.1 2025-12-04T09:31:44.2662045Z - Magma 2.9.0 2025-12-04T09:31:44.2663670Z - Build settings: BLAS_INFO=mkl, BUILD_TYPE=Release, COMMIT_SHA=35b7a9a26c5923d98aebaa41a031dae21788a9ee, CXX_COMPILER=/opt/cache/bin/c++, CXX_FLAGS= -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOCUPTI -DLIBKINETO_NOXPUPTI=ON -DUSE_FBGEMM -DUSE_FBGEMM_GENAI -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -DC10_NODEPRECATED -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=range-loop-construct -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-unknown-pragmas -Wno-unused-parameter -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wsuggest-override -Wno-psabi -Wno-error=old-style-cast -faligned-new -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Wno-stringop-overflow, LAPACK_INFO=mkl, PERF_WITH_AVX=1, PERF_WITH_AVX2=1, TORCH_VERSION=2.10.0, USE_CUDA=OFF, USE_CUDNN=OFF, USE_CUSPARSELT=OFF, USE_GFLAGS=OFF, USE_GLOG=OFF, USE_GLOO=ON, USE_MKL=ON, USE_MKLDNN=ON, USE_MPI=OFF, USE_NCCL=ON, USE_NNPACK=ON, USE_OPENMP=ON, USE_ROCM=ON, USE_ROCM_KERNEL_ASSERT=OFF, USE_XCCL=OFF, USE_XPU=OFF, 2025-12-04T09:31:44.2665714Z 2025-12-04T09:31:44.5065492Z + cd test 2025-12-04T09:31:44.5065859Z + python -c 'import torch; print(torch.__config__.parallel_info())' 2025-12-04T09:31:45.1788627Z ATen/Parallel: 2025-12-04T09:31:45.1788982Z at::get_num_threads() : 128 2025-12-04T09:31:45.1789237Z at::get_num_interop_threads() : 128 2025-12-04T09:31:45.1789487Z OpenMP 201511 (a.k.a. OpenMP 4.5) 2025-12-04T09:31:45.1789769Z omp_get_max_threads() : 128 2025-12-04T09:31:45.1790213Z Intel(R) oneAPI Math Kernel Library Version 2024.2-Product Build 20240605 for Intel(R) 64 architecture applications 2025-12-04T09:31:45.1790643Z mkl_get_max_threads() : 128 2025-12-04T09:31:45.1790942Z Intel(R) MKL-DNN v3.7.1 (Git Hash 8d263e693366ef8db40acc569cc7d8edf644556d) 2025-12-04T09:31:45.1791267Z std::thread::hardware_concurrency() : 128 2025-12-04T09:31:45.1791509Z Environment variables: 2025-12-04T09:31:45.1791709Z OMP_NUM_THREADS : [not set] 2025-12-04T09:31:45.1791912Z MKL_NUM_THREADS : [not set] 2025-12-04T09:31:45.1792125Z ATen parallel backend: OpenMP 2025-12-04T09:31:45.1792265Z 2025-12-04T09:31:45.4106533Z + [[ default == *numpy_2* ]] 2025-12-04T09:31:45.4107002Z + [[ linux-jammy-rocm-py3.10 == *aarch64* ]] 2025-12-04T09:31:45.4107220Z + [[ default == *backward* ]] 2025-12-04T09:31:45.4107424Z + [[ default == *libtorch_agnostic_targetting* ]] 2025-12-04T09:31:45.4107648Z + [[ default == *xla* ]] 2025-12-04T09:31:45.4111042Z + [[ default == *vllm* ]] 2025-12-04T09:31:45.4111224Z + [[ default == *executorch* ]] 2025-12-04T09:31:45.4111410Z + [[ default == \j\i\t\_\l\e\g\a\c\y ]] 2025-12-04T09:31:45.4111598Z + [[ default == \q\u\a\n\t\i\z\a\t\i\o\n ]] 2025-12-04T09:31:45.4111796Z + [[ linux-jammy-rocm-py3.10 == *libtorch* ]] 2025-12-04T09:31:45.4111983Z + [[ default == distributed ]] 2025-12-04T09:31:45.4112159Z + [[ default == *operator_benchmark* ]] 2025-12-04T09:31:45.4112353Z + [[ default == *operator_microbenchmark* ]] 2025-12-04T09:31:45.4112547Z + [[ default == *attention_microbenchmark* ]] 2025-12-04T09:31:45.4112745Z + [[ default == *inductor_distributed* ]] 2025-12-04T09:31:45.4112935Z + [[ default == *inductor-halide* ]] 2025-12-04T09:31:45.4113108Z + [[ default == *inductor-pallas* ]] 2025-12-04T09:31:45.4113303Z + [[ default == *inductor-triton-cpu* ]] 2025-12-04T09:31:45.4113499Z + [[ default == *inductor-micro-benchmark* ]] 2025-12-04T09:31:45.4113707Z + [[ default == *aoti_cross_compile_for_windows* ]] 2025-12-04T09:31:45.4113918Z + [[ default == *huggingface* ]] 2025-12-04T09:31:45.4114101Z + [[ default == *timm* ]] 2025-12-04T09:31:45.4114272Z + [[ default == cachebench ]] 2025-12-04T09:31:45.4114437Z + [[ default == verify_cachebench ]] 2025-12-04T09:31:45.4114626Z + [[ default == *torchbench* ]] 2025-12-04T09:31:45.4114798Z + [[ default == *inductor_cpp_wrapper* ]] 2025-12-04T09:31:45.4114991Z + [[ default == *inductor_core* ]] 2025-12-04T09:31:45.4115170Z + [[ default == *inductor* ]] 2025-12-04T09:31:45.4115340Z + [[ default == *einops* ]] 2025-12-04T09:31:45.4115504Z + [[ default == *dynamo_core* ]] 2025-12-04T09:31:45.4115677Z + [[ default == *dynamo_wrapped* ]] 2025-12-04T09:31:45.4115877Z + [[ linux-jammy-rocm-py3.10 == *rocm* ]] 2025-12-04T09:31:45.4116070Z + [[ -n '' ]] 2025-12-04T09:31:45.4116203Z + [[ 4 == 1 ]] 2025-12-04T09:31:45.4116333Z + [[ 4 == 2 ]] 2025-12-04T09:31:45.4116471Z + [[ 4 -gt 2 ]] 2025-12-04T09:31:45.4116613Z + install_torchvision 2025-12-04T09:31:45.4116808Z + local orig_preload 2025-12-04T09:31:45.4116958Z + local commit 2025-12-04T09:31:45.4121306Z ++ get_pinned_commit vision 2025-12-04T09:31:45.4121593Z ++ cat .github/ci_commit_pins/vision.txt 2025-12-04T09:31:45.4134165Z + commit=617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:31:45.4134334Z + orig_preload= 2025-12-04T09:31:45.4134447Z + '[' -n '' ']' 2025-12-04T09:31:45.4134572Z + [[ linux-jammy-rocm-py3.10 == *cuda* ]] 2025-12-04T09:31:45.4134883Z + pip_build_and_install git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e dist/vision 2025-12-04T09:31:45.4135272Z + local build_target=git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:31:45.4135520Z + local wheel_dir=dist/vision 2025-12-04T09:31:45.4135650Z + local found_whl=0 2025-12-04T09:31:45.4135776Z + for file in "${wheel_dir}"/*.whl 2025-12-04T09:31:45.4135914Z + [[ -f dist/vision/*.whl ]] 2025-12-04T09:31:45.4136042Z + '[' 0 == 0 ']' 2025-12-04T09:31:45.4136341Z + python3 -m pip wheel --no-build-isolation --no-deps -w dist/vision git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:31:45.5649355Z Collecting git+https://github.com/pytorch/vision.git@617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:31:45.5650937Z Cloning https://github.com/pytorch/vision.git (to revision 617079d944b0e72632311c30ae2bbdf1168b901e) to /tmp/pip-req-build-lh1xpfu6 2025-12-04T09:31:45.5674471Z Running command git clone --filter=blob:none --quiet https://github.com/pytorch/vision.git /tmp/pip-req-build-lh1xpfu6 2025-12-04T09:31:48.9428687Z Running command git rev-parse -q --verify 'sha^617079d944b0e72632311c30ae2bbdf1168b901e' 2025-12-04T09:31:48.9441300Z Running command git fetch -q https://github.com/pytorch/vision.git 617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:31:49.5977983Z Resolved https://github.com/pytorch/vision.git to commit 617079d944b0e72632311c30ae2bbdf1168b901e 2025-12-04T09:31:52.1903811Z Preparing metadata (pyproject.toml) ... [?25l- \ | done 2025-12-04T09:31:52.1932255Z [?25hBuilding wheels for collected packages: torchvision 2025-12-04T09:33:10.4063999Z Building wheel for torchvision (pyproject.toml) ... [?25l- \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | / - \ | done 2025-12-04T09:33:10.4080094Z [?25h Created wheel for torchvision: filename=torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl size=1809044 sha256=3c66b5462126300d78e87ac7e0856d60a45439a694c6d38ec6bd922e7334d54a 2025-12-04T09:33:10.4083437Z Stored in directory: /var/lib/jenkins/.cache/pip/wheels/12/b2/29/1f82685c5b5173629e1f36a9b93989ce92ce563e5fb91d27ac 2025-12-04T09:33:10.4101675Z Successfully built torchvision 2025-12-04T09:33:10.4641531Z + for file in "${wheel_dir}"/*.whl 2025-12-04T09:33:10.4641938Z + pip_install_whl dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:33:10.4642422Z + args=('dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl') 2025-12-04T09:33:10.4642720Z + local args 2025-12-04T09:33:10.4642996Z + [[ dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl == *\ * ]] 2025-12-04T09:33:10.4643314Z + for path in "${args[@]}" 2025-12-04T09:33:10.4643621Z + echo 'Installing dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl' 2025-12-04T09:33:10.4644035Z Installing dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:33:10.4644529Z + python3 -mpip install --no-index --no-deps dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:33:10.6209761Z Processing ./dist/vision/torchvision-0.25.0a0+617079d-cp310-cp310-linux_x86_64.whl 2025-12-04T09:33:10.6260313Z Installing collected packages: torchvision 2025-12-04T09:33:10.8445504Z Successfully installed torchvision-0.25.0a0+617079d 2025-12-04T09:33:10.8701660Z + '[' -n '' ']' 2025-12-04T09:33:10.8701826Z + test_python_shard 4 2025-12-04T09:33:10.8701987Z + [[ -z 6 ]] 2025-12-04T09:33:10.8702367Z + python test/run_test.py --exclude-jit-executor --exclude-distributed-tests --exclude-quantization-tests --shard 4 6 --verbose --upload-artifacts-while-running 2025-12-04T09:33:12.7563688Z Excluding inductor/test_max_autotune on ROCm 2025-12-04T09:33:12.7564175Z Excluding test_cuda_nvml_based_avail on ROCm 2025-12-04T09:33:14.0335673Z Downloading https://ossci-metrics.s3.amazonaws.com/disabled-tests-condensed.json to /var/lib/jenkins/pytorch/test/.pytorch-disabled-tests.json 2025-12-04T09:33:14.3986460Z Ignoring disabled issues: [''] 2025-12-04T09:33:14.4032490Z Found test times from artifacts 2025-12-04T09:33:14.4522044Z Found test times from artifacts 2025-12-04T09:33:14.4530412Z Running all tests 2025-12-04T09:33:14.4822699Z Running parallel tests on 1 processes 2025-12-04T09:33:14.4826340Z Name: tests to run (est. time: 180.8min) 2025-12-04T09:33:14.4826578Z Serial tests (98): 2025-12-04T09:33:14.4826829Z inductor/test_aot_inductor 3/3 2025-12-04T09:33:14.4827111Z inductor/test_torchinductor_codegen_dynamic_shapes 2/4 2025-12-04T09:33:14.4827470Z inductor/test_torchinductor_opinfo 3/12 2025-12-04T09:33:14.4827729Z inductor/test_torchinductor_opinfo 9/12 2025-12-04T09:33:14.4828021Z dynamo/test_dynamic_shapes 1/2 2025-12-04T09:33:14.4828264Z inductor/test_cpu_repro 5/5 2025-12-04T09:33:14.4828530Z inductor/test_inductor_freezing 1/1 2025-12-04T09:33:14.4828756Z inductor/test_flex_attention 2/4 2025-12-04T09:33:14.4829025Z inductor/test_cpu_cpp_wrapper 1/1 2025-12-04T09:33:14.4829256Z inductor/test_compile_subprocess 3/3 2025-12-04T09:33:14.4829556Z inductor/test_multi_kernel 1/1 2025-12-04T09:33:14.4829767Z inductor/test_analysis 1/1 2025-12-04T09:33:14.4830153Z inductor/test_pad_mm 1/1 2025-12-04T09:33:14.4830375Z inductor/test_triton_syntax 1/1 2025-12-04T09:33:14.4830633Z test_sparse_semi_structured 1/1 2025-12-04T09:33:14.4830928Z inductor/test_op_completeness 1/1 2025-12-04T09:33:14.4834331Z inductor/test_subgraph_choice 1/1 2025-12-04T09:33:14.4834531Z inductor/test_b2b_gemm 1/1 2025-12-04T09:33:14.4834730Z inductor/test_triton_heuristics 1/1 2025-12-04T09:33:14.4834928Z inductor/test_ck_backend 1/1 2025-12-04T09:33:14.4837200Z test_testing 1/1 2025-12-04T09:33:14.4837439Z inductor/test_autoheuristic 1/1 2025-12-04T09:33:14.4837686Z export/test_strict_export_v2 1/1 2025-12-04T09:33:14.4837884Z inductor/test_segmented_tree 1/1 2025-12-04T09:33:14.4838093Z inductor/test_kernel_optimization 1/1 2025-12-04T09:33:14.4838311Z inductor/test_inductor_scheduler 1/1 2025-12-04T09:33:14.4838475Z inductor/test_remote_cache 1/1 2025-12-04T09:33:14.4838642Z inductor/test_coordinate_descent_tuner 1/1 2025-12-04T09:33:14.4838816Z inductor/test_inplace_padding 1/1 2025-12-04T09:33:14.4838973Z inductor/test_minifier_utils 1/1 2025-12-04T09:33:14.4839122Z inductor/test_debug_trace 1/1 2025-12-04T09:33:14.4839270Z dynamo/test_recompiles 1/1 2025-12-04T09:33:14.4839416Z inductor/test_foreach 1/1 2025-12-04T09:33:14.4839566Z dynamo/test_buffers_override 1/1 2025-12-04T09:33:14.4839719Z inductor/test_inplacing_pass 1/1 2025-12-04T09:33:14.4839879Z inductor/test_aot_inductor_custom_ops 1/1 2025-12-04T09:33:14.4840053Z inductor/test_split_cat_fx_passes 1/1 2025-12-04T09:33:14.4840211Z export/test_cpp_serdes 1/1 2025-12-04T09:33:14.4840356Z dynamo/test_pgo 1/1 2025-12-04T09:33:14.4840485Z dynamo/test_config 1/1 2025-12-04T09:33:14.4840627Z dynamo/test_metrics_context 1/1 2025-12-04T09:33:14.4840777Z export/test_package 1/1 2025-12-04T09:33:14.4840921Z export/test_export_opinfo 1/1 2025-12-04T09:33:14.4841064Z dynamo/test_nops 1/1 2025-12-04T09:33:14.4841210Z inductor/test_graph_transform_observer 1/1 2025-12-04T09:33:14.4841377Z inductor/test_memory 1/1 2025-12-04T09:33:14.4841515Z inductor/test_control_flow 1/4 2025-12-04T09:33:14.4841664Z inductor/test_pallas 1/1 2025-12-04T09:33:14.4841811Z dynamo/test_callback 1/1 2025-12-04T09:33:14.4841969Z dynamo/test_cudagraphs_expandable_segments 1/1 2025-12-04T09:33:14.4842306Z dynamo/test_aot_autograd_cache 1/1 2025-12-04T09:33:14.4842462Z test_utils_config_module 1/1 2025-12-04T09:33:14.4842600Z test_hop_infra 1/1 2025-12-04T09:33:14.4842730Z test_transformers 1/1 2025-12-04T09:33:14.4842890Z torch_np/numpy_tests/core/test_scalarinherit 1/1 2025-12-04T09:33:14.4843070Z test_tensorexpr_pybind 1/1 2025-12-04T09:33:14.4843206Z test_fx_experimental 1/1 2025-12-04T09:33:14.4843347Z test_namedtensor 1/1 2025-12-04T09:33:14.4843483Z test_dataloader 1/1 2025-12-04T09:33:14.4843606Z test_decomp 6/12 2025-12-04T09:33:14.4843739Z test_decomp 12/12 2025-12-04T09:33:14.4843866Z test_ci_sanity_check_fail 1/1 2025-12-04T09:33:14.4844013Z test_meta 3/3 2025-12-04T09:33:14.4844140Z test_ops_gradients 2/2 2025-12-04T09:33:14.4844270Z test_linalg 2/2 2025-12-04T09:33:14.4844416Z nn/test_packed_sequence 1/1 2025-12-04T09:33:14.4844566Z test_numa_binding 1/1 2025-12-04T09:33:14.4844701Z test_pruning_op 1/1 2025-12-04T09:33:14.4844831Z test_jit_fuser_te 1/1 2025-12-04T09:33:14.4844965Z nn/test_dropout 1/1 2025-12-04T09:33:14.4845094Z functorch/dim/test_split 1/1 2025-12-04T09:33:14.4845251Z torch_np/numpy_tests/lib/test_type_check 1/1 2025-12-04T09:33:14.4845434Z cpp_extensions/test_libtorch_agnostic 1/1 2025-12-04T09:33:14.4845603Z test_cpp_extensions_stream_and_event 1/1 2025-12-04T09:33:14.4845766Z profiler/test_execution_trace 1/1 2025-12-04T09:33:14.4845916Z test_dispatch 1/1 2025-12-04T09:33:14.4846049Z test_datapipe 1/1 2025-12-04T09:33:14.4846168Z test_ops 1/5 2025-12-04T09:33:14.4846309Z torch_np/numpy_tests/linalg/test_linalg 1/1 2025-12-04T09:33:14.4846466Z nn/test_module_hooks 1/1 2025-12-04T09:33:14.4846619Z torch_np/numpy_tests/lib/test_twodim_base 1/1 2025-12-04T09:33:14.4846852Z profiler/test_memory_profiler 1/1 2025-12-04T09:33:14.4847073Z test_serialization 1/1 2025-12-04T09:33:14.4847207Z test_indexing 1/1 2025-12-04T09:33:14.4847353Z torch_np/numpy_tests/fft/test_pocketfft 1/1 2025-12-04T09:33:14.4847522Z functorch/test_ops 2/4 2025-12-04T09:33:14.4847665Z torch_np/numpy_tests/core/test_dlpack 1/1 2025-12-04T09:33:14.4847838Z test_multiprocessing_spawn 1/1 2025-12-04T09:33:14.4847978Z test_mkldnn 1/1 2025-12-04T09:33:14.4848117Z torch_np/numpy_tests/lib/test_index_tricks 1/1 2025-12-04T09:33:14.4848332Z test_jit_autocast 1/1 2025-12-04T09:33:14.4848442Z nn/test_pooling 1/1 2025-12-04T09:33:14.4848554Z lazy/test_reuse_ir 1/1 2025-12-04T09:33:14.4848673Z test_functional_autograd_benchmark 1/1 2025-12-04T09:33:14.4848872Z cpp_extensions/libtorch_agnostic_2_10_extension/test_version_compatibility 1/1 2025-12-04T09:33:14.4849059Z test_autoload_disable 1/1 2025-12-04T09:33:14.4849185Z test_cpp_extensions_aot_ninja 1/1 2025-12-04T09:33:14.4849321Z test_cpp_extensions_aot_no_ninja 1/1 2025-12-04T09:33:14.4849452Z Parallel tests (0): 2025-12-04T09:33:14.4849567Z Name: excluded (est. time: 0.0min) 2025-12-04T09:33:14.4849687Z Serial tests (0): 2025-12-04T09:33:14.4849785Z Parallel tests (0): 2025-12-04T09:33:14.4849954Z Running inductor/test_aot_inductor 3/3 ... [2025-12-04 09:33:14.482735][2224618.940056434] 2025-12-04T09:33:14.4850150Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:33:14.4850624Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor.py', '--shard-id=3', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:33:14.483024] 2025-12-04T09:45:06.4964802Z 2025-12-04T09:45:06.4965961Z PRINTING LOG FILE of inductor/test_aot_inductor 3/3 (test/test-reports/inductor.test_aot_inductor_3.3_f0c7cbe1a17c7fbf_.log) 2025-12-04T09:45:06.4967304Z Test results will be stored in test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:45:06.4971958Z ============================= test session starts ============================== 2025-12-04T09:45:06.4972596Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:45:06.4973173Z cachedir: .pytest_cache 2025-12-04T09:45:06.4973803Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:45:06.4974385Z rootdir: /var/lib/jenkins/pytorch 2025-12-04T09:45:06.4974617Z configfile: pytest.ini 2025-12-04T09:45:06.4975065Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:45:06.4975540Z collecting ... collected 934 items 2025-12-04T09:45:06.4975820Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:45:06.5026368Z Running 150 items in this shard: test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda, test/inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda 2025-12-04T09:45:06.5047883Z 2025-12-04T09:45:06.5048253Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:33:23.488000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5048761Z W1204 09:33:23.489000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5049004Z PASSED [12.7980s] [ 0%] 2025-12-04T09:45:06.5049287Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [4.8607s] [ 1%] 2025-12-04T09:45:06.5049828Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py W1204 09:33:43.020000 9298 site-packages/torch/_export/__init__.py:71] +============================+ 2025-12-04T09:45:06.5050420Z W1204 09:33:43.020000 9298 site-packages/torch/_export/__init__.py:72] | !!! WARNING !!! | 2025-12-04T09:45:06.5050681Z W1204 09:33:43.020000 9298 site-packages/torch/_export/__init__.py:73] +============================+ 2025-12-04T09:45:06.5051194Z W1204 09:33:43.021000 9298 site-packages/torch/_export/__init__.py:74] torch._export.aot_compile()/torch._export.aot_load() is being deprecated, please switch to directly calling torch._inductor.aoti_compile_and_package(torch.export.export())/torch._inductor.aoti_load_package() instead. 2025-12-04T09:45:06.5051631Z PASSED [2.8114s] [ 2%] 2025-12-04T09:45:06.5052025Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:33:43.151000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5052503Z W1204 09:33:43.152000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5052706Z PASSED [2.9376s] [ 2%] 2025-12-04T09:45:06.5053095Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:33:46.367000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5053607Z W1204 09:33:46.368000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5053810Z PASSED [5.0670s] [ 2%] 2025-12-04T09:45:06.5054190Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:33:51.154000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5054659Z W1204 09:33:51.154000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5054854Z FAILED [6.3382s] [ 2%] 2025-12-04T09:45:06.5055233Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:33:57.495000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5055702Z W1204 09:33:57.495000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5055901Z FAILED [3.0370s] [ 2%] 2025-12-04T09:45:06.5056282Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:00.535000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5056800Z W1204 09:34:00.536000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5056998Z FAILED [3.0923s] [ 2%] 2025-12-04T09:45:06.5057378Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:03.627000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5057845Z W1204 09:34:03.627000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5058041Z FAILED [5.8673s] [ 2%] 2025-12-04T09:45:06.5058424Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:09.494000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5058931Z W1204 09:34:09.495000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5059127Z FAILED [3.0095s] [ 2%] 2025-12-04T09:45:06.5059510Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:12.506000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5059984Z W1204 09:34:12.507000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5060187Z FAILED [3.0399s] [ 2%] 2025-12-04T09:45:06.5060574Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:15.547000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5061055Z W1204 09:34:15.548000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5061258Z FAILED [3.0553s] [ 2%] 2025-12-04T09:45:06.5061645Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:18.608000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5062124Z W1204 09:34:18.608000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5062329Z FAILED [3.0247s] [ 2%] 2025-12-04T09:45:06.5062754Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:21.630000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5063230Z W1204 09:34:21.630000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5063436Z FAILED [3.0609s] [ 2%] 2025-12-04T09:45:06.5063822Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:24.693000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5064305Z W1204 09:34:24.694000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5064508Z FAILED [8.7884s] [ 2%] 2025-12-04T09:45:06.5064894Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:33.482000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5065371Z W1204 09:34:33.482000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5065583Z FAILED [8.1315s] [ 2%] 2025-12-04T09:45:06.5065967Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:41.617000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5066440Z W1204 09:34:41.617000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5066644Z FAILED [8.0859s] [ 2%] 2025-12-04T09:45:06.5067071Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:49.703000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5067542Z W1204 09:34:49.703000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5067742Z FAILED [4.5393s] [ 2%] 2025-12-04T09:45:06.5068121Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:34:54.243000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5068621Z W1204 09:34:54.243000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5068819Z FAILED [6.5972s] [ 2%] 2025-12-04T09:45:06.5069197Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:35:00.845000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5069672Z W1204 09:35:00.846000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5069873Z PASSED [11.6559s] [ 2%] 2025-12-04T09:45:06.5070256Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:35:12.499000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5070727Z W1204 09:35:12.499000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5070926Z FAILED [8.7425s] [ 2%] 2025-12-04T09:45:06.5071305Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:35:21.245000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5071774Z W1204 09:35:21.245000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5071972Z FAILED [6.4504s] [ 2%] 2025-12-04T09:45:06.5072446Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:35:27.698000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5072921Z W1204 09:35:27.698000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5073121Z FAILED [3.1455s] [ 2%] 2025-12-04T09:45:06.5073502Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:35:30.843000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5073973Z W1204 09:35:30.843000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5074173Z FAILED [3.0648s] [ 2%] 2025-12-04T09:45:06.5074557Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:35:33.910000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5075025Z W1204 09:35:33.911000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5075227Z FAILED [3.0759s] [ 2%] 2025-12-04T09:45:06.5075621Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:35:36.992000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5076089Z W1204 09:35:36.993000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5076279Z FAILED [3.0232s] [ 2%] 2025-12-04T09:45:06.5076650Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:35:40.015000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5077154Z W1204 09:35:40.015000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5077380Z FAILED [6.5313s] [ 2%] 2025-12-04T09:45:06.5077752Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:35:46.545000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5078215Z W1204 09:35:46.546000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5078410Z FAILED [3.1357s] [ 2%] 2025-12-04T09:45:06.5078783Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:35:49.681000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5079249Z W1204 09:35:49.682000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5079442Z FAILED [4.5588s] [ 2%] 2025-12-04T09:45:06.5079818Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:35:54.243000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5080285Z W1204 09:35:54.243000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5080477Z FAILED [6.6000s] [ 2%] 2025-12-04T09:45:06.5080864Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:36:00.844000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5081328Z W1204 09:36:00.844000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5081520Z FAILED [3.0815s] [ 2%] 2025-12-04T09:45:06.5081928Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:36:03.927000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5082392Z W1204 09:36:03.927000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5082583Z FAILED [4.6693s] [ 2%] 2025-12-04T09:45:06.5082956Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:36:08.597000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5083417Z W1204 09:36:08.598000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5083607Z FAILED [7.3739s] [ 2%] 2025-12-04T09:45:06.5083984Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:36:15.975000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5084448Z W1204 09:36:15.975000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5084640Z FAILED [9.1520s] [ 2%] 2025-12-04T09:45:06.5085010Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:36:25.127000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5085471Z W1204 09:36:25.127000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5085661Z FAILED [3.0411s] [ 2%] 2025-12-04T09:45:06.5086031Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:36:28.169000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5086498Z W1204 09:36:28.169000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5086787Z FAILED [4.7527s] [ 2%] 2025-12-04T09:45:06.5087161Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:36:32.923000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5087623Z W1204 09:36:32.924000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5087818Z FAILED [12.3107s] [ 2%] 2025-12-04T09:45:06.5088195Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:36:45.235000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5088659Z W1204 09:36:45.236000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5088853Z FAILED [8.0937s] [ 2%] 2025-12-04T09:45:06.5089227Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:36:53.331000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5089688Z W1204 09:36:53.332000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5089880Z FAILED [12.2574s] [ 2%] 2025-12-04T09:45:06.5090257Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:37:05.589000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5090720Z W1204 09:37:05.589000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5090912Z FAILED [4.4793s] [ 2%] 2025-12-04T09:45:06.5091316Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:37:10.071000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5091783Z W1204 09:37:10.071000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5091974Z FAILED [8.9116s] [ 2%] 2025-12-04T09:45:06.5092348Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:37:18.984000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5092810Z W1204 09:37:18.984000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5093001Z PASSED [7.7889s] [ 2%] 2025-12-04T09:45:06.5093378Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:37:26.773000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5093848Z W1204 09:37:26.774000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5094040Z FAILED [6.4163s] [ 2%] 2025-12-04T09:45:06.5094414Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:37:33.194000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5094874Z W1204 09:37:33.194000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5095069Z FAILED [3.0219s] [ 2%] 2025-12-04T09:45:06.5095442Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:37:36.217000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5095939Z W1204 09:37:36.217000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5096130Z FAILED [8.4732s] [ 2%] 2025-12-04T09:45:06.5096503Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:37:45.054000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5096995Z W1204 09:37:45.054000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5097189Z PASSED [3.3638s] [ 2%] 2025-12-04T09:45:06.5097564Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:37:48.061000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5098032Z W1204 09:37:48.061000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5098228Z PASSED [2.9953s] [ 2%] 2025-12-04T09:45:06.5098605Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:37:51.052000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5099070Z W1204 09:37:51.052000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5099264Z PASSED [2.9973s] [ 2%] 2025-12-04T09:45:06.5099644Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:37:54.052000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5100149Z W1204 09:37:54.052000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5100342Z PASSED [2.9983s] [ 2%] 2025-12-04T09:45:06.5100718Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:37:57.050000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5101181Z W1204 09:37:57.051000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5101372Z PASSED [3.0454s] [ 2%] 2025-12-04T09:45:06.5101745Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:38:00.100000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5102209Z W1204 09:38:00.101000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5102400Z FAILED [3.0237s] [ 2%] 2025-12-04T09:45:06.5102777Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda <- test/inductor/test_torchinductor.py W1204 09:38:03.123000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5103243Z W1204 09:38:03.123000 9298 site-packages/torch/_inductor/utils.py:2565] [0/0] DeviceCopy in input program 2025-12-04T09:45:06.5103435Z FAILED [3.0111s] [ 2%] 2025-12-04T09:45:06.5103698Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.9136s] [ 2%] 2025-12-04T09:45:06.5104134Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [5.0526s] [ 2%] 2025-12-04T09:45:06.5104568Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7574s] [ 2%] 2025-12-04T09:45:06.5105006Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.9136s] [ 2%] 2025-12-04T09:45:06.5105473Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [4.8971s] [ 2%] 2025-12-04T09:45:06.5105906Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7630s] [ 2%] 2025-12-04T09:45:06.5106338Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7863s] [ 2%] 2025-12-04T09:45:06.5106867Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7862s] [ 2%] 2025-12-04T09:45:06.5107312Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7910s] [ 2%] 2025-12-04T09:45:06.5107754Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7579s] [ 2%] 2025-12-04T09:45:06.5108191Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7306s] [ 2%] 2025-12-04T09:45:06.5108635Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7327s] [ 2%] 2025-12-04T09:45:06.5109072Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.8052s] [ 2%] 2025-12-04T09:45:06.5109505Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7383s] [ 2%] 2025-12-04T09:45:06.5109968Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7056s] [ 2%] 2025-12-04T09:45:06.5110404Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7894s] [ 2%] 2025-12-04T09:45:06.5110836Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.8455s] [ 2%] 2025-12-04T09:45:06.5111268Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [3.1521s] [ 2%] 2025-12-04T09:45:06.5111702Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.8288s] [ 2%] 2025-12-04T09:45:06.5112140Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.9504s] [ 2%] 2025-12-04T09:45:06.5112571Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.8230s] [ 2%] 2025-12-04T09:45:06.5113009Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7971s] [ 2%] 2025-12-04T09:45:06.5113465Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7833s] [ 2%] 2025-12-04T09:45:06.5113981Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7999s] [ 2%] 2025-12-04T09:45:06.5114413Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.8258s] [ 2%] 2025-12-04T09:45:06.5114846Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [3.9044s] [ 2%] 2025-12-04T09:45:06.5115313Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.8248s] [ 2%] 2025-12-04T09:45:06.5115748Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7718s] [ 2%] 2025-12-04T09:45:06.5116182Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7857s] [ 2%] 2025-12-04T09:45:06.5116619Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.8156s] [ 2%] 2025-12-04T09:45:06.5117124Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7511s] [ 2%] 2025-12-04T09:45:06.5117566Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.6872s] [ 2%] 2025-12-04T09:45:06.5118007Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.6799s] [ 2%] 2025-12-04T09:45:06.5118445Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [5.8516s] [ 2%] 2025-12-04T09:45:06.5118888Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [7.1998s] [ 2%] 2025-12-04T09:45:06.5119330Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [7.5390s] [ 2%] 2025-12-04T09:45:06.5119812Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [4.8328s] [ 2%] 2025-12-04T09:45:06.5120257Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [5.3090s] [ 2%] 2025-12-04T09:45:06.5120700Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [4.8998s] [ 2%] 2025-12-04T09:45:06.5121145Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7988s] [ 2%] 2025-12-04T09:45:06.5121587Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [5.0503s] [ 2%] 2025-12-04T09:45:06.5122029Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.9774s] [ 2%] 2025-12-04T09:45:06.5122467Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.9146s] [ 2%] 2025-12-04T09:45:06.5122910Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.6925s] [ 2%] 2025-12-04T09:45:06.5123350Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.6964s] [ 2%] 2025-12-04T09:45:06.5123791Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [8.2638s] [ 2%] 2025-12-04T09:45:06.5124237Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [3.1291s] [ 2%] 2025-12-04T09:45:06.5124683Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.8826s] [ 2%] 2025-12-04T09:45:06.5125127Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_repeated_calling_cuda <- test/inductor/test_torchinductor.py PASSED [2.7514s] [ 2%] 2025-12-04T09:45:06.5125602Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6608s] [ 2%] 2025-12-04T09:45:06.5126037Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [5.2091s] [ 2%] 2025-12-04T09:45:06.5126472Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [3.1183s] [ 2%] 2025-12-04T09:45:06.5126946Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.7012s] [ 2%] 2025-12-04T09:45:06.5127382Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.9438s] [ 2%] 2025-12-04T09:45:06.5127816Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [4.7233s] [ 2%] 2025-12-04T09:45:06.5128251Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6181s] [ 2%] 2025-12-04T09:45:06.5128686Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6765s] [ 2%] 2025-12-04T09:45:06.5129122Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.5938s] [ 2%] 2025-12-04T09:45:06.5129557Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [5.0139s] [ 2%] 2025-12-04T09:45:06.5130029Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.5414s] [ 2%] 2025-12-04T09:45:06.5130464Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [3.1327s] [ 2%] 2025-12-04T09:45:06.5130903Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [7.7995s] [ 2%] 2025-12-04T09:45:06.5131337Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.5222s] [ 2%] 2025-12-04T09:45:06.5131776Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.4949s] [ 2%] 2025-12-04T09:45:06.5132212Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.7725s] [ 2%] 2025-12-04T09:45:06.5132639Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.4877s] [ 2%] 2025-12-04T09:45:06.5133069Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.6245s] [ 2%] 2025-12-04T09:45:06.5133496Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.4328s] [ 2%] 2025-12-04T09:45:06.5133924Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.5176s] [ 2%] 2025-12-04T09:45:06.5134349Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [7.2477s] [ 2%] 2025-12-04T09:45:06.5134775Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.5848s] [ 2%] 2025-12-04T09:45:06.5135240Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.3631s] [ 2%] 2025-12-04T09:45:06.5135671Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.1769s] [ 2%] 2025-12-04T09:45:06.5136103Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.4740s] [ 2%] 2025-12-04T09:45:06.5136537Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [8.4059s] [ 2%] 2025-12-04T09:45:06.5137010Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [7.7817s] [ 2%] 2025-12-04T09:45:06.5137447Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6326s] [ 2%] 2025-12-04T09:45:06.5137883Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.5359s] [ 2%] 2025-12-04T09:45:06.5138318Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.5674s] [ 2%] 2025-12-04T09:45:06.5138751Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6744s] [ 2%] 2025-12-04T09:45:06.5139185Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6511s] [ 2%] 2025-12-04T09:45:06.5139653Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [6.2425s] [ 2%] 2025-12-04T09:45:06.5140090Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [5.7349s] [ 2%] 2025-12-04T09:45:06.5140522Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6917s] [ 2%] 2025-12-04T09:45:06.5140956Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6572s] [ 2%] 2025-12-04T09:45:06.5141390Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [3.0685s] [ 2%] 2025-12-04T09:45:06.5141820Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [6.0713s] [ 2%] 2025-12-04T09:45:06.5142260Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [5.1367s] [ 2%] 2025-12-04T09:45:06.5142688Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [5.6418s] [ 2%] 2025-12-04T09:45:06.5143119Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [5.3283s] [ 2%] 2025-12-04T09:45:06.5143542Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.7493s] [ 2%] 2025-12-04T09:45:06.5143965Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6770s] [ 2%] 2025-12-04T09:45:06.5144391Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.7121s] [ 2%] 2025-12-04T09:45:06.5144817Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6752s] [ 2%] 2025-12-04T09:45:06.5145275Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6704s] [ 2%] 2025-12-04T09:45:06.5145699Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.6678s] [ 2%] 2025-12-04T09:45:06.5146124Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.9040s] [ 2%] 2025-12-04T09:45:06.5146547Z inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_upper_bound_i64_cuda <- test/inductor/test_torchinductor.py PASSED [2.8521s] [ 2%] 2025-12-04T09:45:06.5146835Z 2025-12-04T09:45:06.5146902Z =================================== FAILURES =================================== 2025-12-04T09:45:06.5147107Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5147299Z Traceback (most recent call last): 2025-12-04T09:45:06.5147509Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5147704Z return value(self) 2025-12-04T09:45:06.5147918Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5148155Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5148386Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5148621Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5148875Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5149186Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5149355Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5149447Z 2025-12-04T09:45:06.5149495Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5149677Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5149918Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5150056Z 2025-12-04T09:45:06.5150133Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5150435Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5150655Z 2025-12-04T09:45:06.5150746Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5150952Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5151112Z unimplemented [] 2025-12-04T09:45:06.5151229Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5151373Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5151483Z graph_break [] 2025-12-04T09:45:06.5151616Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5151965Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5152282Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5152434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5152584Z unimplemented [] 2025-12-04T09:45:06.5152698Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5152832Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5152941Z graph_break [] 2025-12-04T09:45:06.5153074Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5153416Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5153767Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5153915Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5154065Z unimplemented [] 2025-12-04T09:45:06.5154178Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5154314Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5154421Z graph_break [] 2025-12-04T09:45:06.5154548Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5154887Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5155197Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5155346Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5155498Z unimplemented [] 2025-12-04T09:45:06.5155611Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5155745Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5155852Z graph_break [] 2025-12-04T09:45:06.5155979Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5156316Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5156624Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5156848Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5157037Z Traceback (most recent call last): 2025-12-04T09:45:06.5157235Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5157473Z return value(self) 2025-12-04T09:45:06.5157691Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5157929Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5158155Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5158393Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5158649Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5158917Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5159080Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5159168Z 2025-12-04T09:45:06.5159217Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5159401Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5159747Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5159888Z 2025-12-04T09:45:06.5159964Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5160257Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5160475Z 2025-12-04T09:45:06.5160564Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5160763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5160913Z unimplemented [] 2025-12-04T09:45:06.5161029Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5161165Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5161273Z graph_break [] 2025-12-04T09:45:06.5161403Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5161758Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5162107Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5162264Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5162422Z unimplemented [] 2025-12-04T09:45:06.5162543Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5162682Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5162796Z graph_break [] 2025-12-04T09:45:06.5162932Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5163281Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5163598Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5163753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5163913Z unimplemented [] 2025-12-04T09:45:06.5164033Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5164175Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5164289Z graph_break [] 2025-12-04T09:45:06.5164425Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5164772Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5165085Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5165238Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5165394Z unimplemented [] 2025-12-04T09:45:06.5165514Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5165655Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5165804Z graph_break [] 2025-12-04T09:45:06.5165939Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5166289Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5166604Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5166796Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5166954Z unimplemented [] 2025-12-04T09:45:06.5167072Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5167213Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5167327Z graph_break [] 2025-12-04T09:45:06.5167461Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5167810Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5168125Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5168308Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5168503Z Traceback (most recent call last): 2025-12-04T09:45:06.5168706Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5168905Z return value(self) 2025-12-04T09:45:06.5169122Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5169361Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5169590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5169831Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5170094Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5170406Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5170575Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5170664Z 2025-12-04T09:45:06.5170720Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5170904Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5171149Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5171290Z 2025-12-04T09:45:06.5171370Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5171668Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5171890Z 2025-12-04T09:45:06.5171984Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5172190Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5172356Z unimplemented [] 2025-12-04T09:45:06.5172477Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5172618Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5172732Z graph_break [] 2025-12-04T09:45:06.5172868Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5173222Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5173540Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5173695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5173855Z unimplemented [] 2025-12-04T09:45:06.5173974Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5174116Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5174232Z graph_break [] 2025-12-04T09:45:06.5174402Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5174754Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5175072Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5175225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5175382Z unimplemented [] 2025-12-04T09:45:06.5175504Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5175645Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5175760Z graph_break [] 2025-12-04T09:45:06.5175895Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5176242Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5176557Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5176709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5176896Z unimplemented [] 2025-12-04T09:45:06.5177015Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5177156Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5177272Z graph_break [] 2025-12-04T09:45:06.5177406Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5177754Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5178065Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5178216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5178370Z unimplemented [] 2025-12-04T09:45:06.5178493Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5178634Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5178790Z graph_break [] 2025-12-04T09:45:06.5178924Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5179268Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5179582Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5179735Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5179888Z unimplemented [] 2025-12-04T09:45:06.5180007Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5180148Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5180262Z graph_break [] 2025-12-04T09:45:06.5180396Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5180744Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5181067Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5181249Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5181446Z Traceback (most recent call last): 2025-12-04T09:45:06.5181649Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5181849Z return value(self) 2025-12-04T09:45:06.5182068Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5182311Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5182545Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5182786Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5183082Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5183361Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5183528Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5183622Z 2025-12-04T09:45:06.5183670Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5183854Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5184098Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5184242Z 2025-12-04T09:45:06.5184317Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5184618Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5184843Z 2025-12-04T09:45:06.5184934Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5185138Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5185295Z unimplemented [] 2025-12-04T09:45:06.5185415Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5185559Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5185673Z graph_break [] 2025-12-04T09:45:06.5185808Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5186156Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5186470Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5186621Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5186810Z unimplemented [] 2025-12-04T09:45:06.5186923Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5187066Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5195313Z graph_break [] 2025-12-04T09:45:06.5195449Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5195794Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5196108Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5196260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5196417Z unimplemented [] 2025-12-04T09:45:06.5196537Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5196680Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5196824Z graph_break [] 2025-12-04T09:45:06.5196957Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5197304Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5197624Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5197771Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5197919Z unimplemented [] 2025-12-04T09:45:06.5198034Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5198169Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5198278Z graph_break [] 2025-12-04T09:45:06.5198405Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5198747Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5199054Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5199199Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5199383Z unimplemented [] 2025-12-04T09:45:06.5199501Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5199637Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5199744Z graph_break [] 2025-12-04T09:45:06.5199874Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5200213Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5200522Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5200669Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5200819Z unimplemented [] 2025-12-04T09:45:06.5200932Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5201068Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5201176Z graph_break [] 2025-12-04T09:45:06.5201310Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5201655Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5201961Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5202107Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5202256Z unimplemented [] 2025-12-04T09:45:06.5202371Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5202506Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5202614Z graph_break [] 2025-12-04T09:45:06.5202741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5203081Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5203395Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5203574Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5203804Z Traceback (most recent call last): 2025-12-04T09:45:06.5203999Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5204192Z return value(self) 2025-12-04T09:45:06.5204404Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5204637Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5204860Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5205091Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5205342Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5205611Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5205773Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5205861Z 2025-12-04T09:45:06.5205906Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5206084Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5206322Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5206461Z 2025-12-04T09:45:06.5206534Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5206862Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5207082Z 2025-12-04T09:45:06.5207171Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5207364Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5207515Z unimplemented [] 2025-12-04T09:45:06.5207661Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5207802Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5207912Z graph_break [] 2025-12-04T09:45:06.5208040Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5208380Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5208692Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5208838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5208993Z unimplemented [] 2025-12-04T09:45:06.5209106Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5209242Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5209350Z graph_break [] 2025-12-04T09:45:06.5209477Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5209821Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5210136Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5210283Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5210433Z unimplemented [] 2025-12-04T09:45:06.5210545Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5210682Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5210793Z graph_break [] 2025-12-04T09:45:06.5210921Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5211263Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5211570Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5211720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5211904Z unimplemented [] 2025-12-04T09:45:06.5212016Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5212152Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5212259Z graph_break [] 2025-12-04T09:45:06.5212384Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5212720Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5213024Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5213169Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5213316Z unimplemented [] 2025-12-04T09:45:06.5213427Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5213560Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5213666Z graph_break [] 2025-12-04T09:45:06.5213793Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5214139Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5214448Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5214598Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5214746Z unimplemented [] 2025-12-04T09:45:06.5214858Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5214992Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5215100Z graph_break [] 2025-12-04T09:45:06.5215226Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5215597Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5215910Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5216055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5216204Z unimplemented [] 2025-12-04T09:45:06.5216315Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5216451Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5216557Z graph_break [] 2025-12-04T09:45:06.5216682Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5217052Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5217358Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5217503Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5217650Z unimplemented [] 2025-12-04T09:45:06.5217765Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5217902Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5218009Z graph_break [] 2025-12-04T09:45:06.5218135Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5218474Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5218783Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5218959Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5219146Z Traceback (most recent call last): 2025-12-04T09:45:06.5219341Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5219534Z return value(self) 2025-12-04T09:45:06.5219748Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5219982Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5220260Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5220504Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5220771Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5221042Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5221205Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5221293Z 2025-12-04T09:45:06.5221341Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5221522Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5221761Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5221898Z 2025-12-04T09:45:06.5221975Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5222269Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5222486Z 2025-12-04T09:45:06.5222578Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5222779Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5222929Z unimplemented [] 2025-12-04T09:45:06.5223042Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5223178Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5223284Z graph_break [] 2025-12-04T09:45:06.5223413Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5223786Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5224096Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5224412Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5224600Z unimplemented [] 2025-12-04T09:45:06.5224761Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5224934Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5225079Z graph_break [] 2025-12-04T09:45:06.5225268Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5225649Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5225990Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5226188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5226376Z unimplemented [] 2025-12-04T09:45:06.5226537Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5226707Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5226887Z graph_break [] 2025-12-04T09:45:06.5227067Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5227447Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5227784Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5227984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5240642Z unimplemented [] 2025-12-04T09:45:06.5240778Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5240933Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5241050Z graph_break [] 2025-12-04T09:45:06.5241189Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5241552Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5241948Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5242103Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5242257Z unimplemented [] 2025-12-04T09:45:06.5242375Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5242510Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5242620Z graph_break [] 2025-12-04T09:45:06.5242753Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5243099Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5243408Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5243560Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5243718Z unimplemented [] 2025-12-04T09:45:06.5243837Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5243973Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5244078Z graph_break [] 2025-12-04T09:45:06.5244208Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5244553Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5244865Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5245014Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5245161Z unimplemented [] 2025-12-04T09:45:06.5245279Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5245416Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5245526Z graph_break [] 2025-12-04T09:45:06.5245688Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5246036Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5246345Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5246494Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5246645Z unimplemented [] 2025-12-04T09:45:06.5246798Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5246930Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5247037Z graph_break [] 2025-12-04T09:45:06.5247166Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5247507Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5247820Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5247970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5248120Z unimplemented [] 2025-12-04T09:45:06.5248234Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5248374Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5248481Z graph_break [] 2025-12-04T09:45:06.5248608Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5248951Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5249257Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5249441Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5249629Z Traceback (most recent call last): 2025-12-04T09:45:06.5249835Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5250065Z return value(self) 2025-12-04T09:45:06.5250283Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5250527Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5250761Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5250999Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5251259Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5251534Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5251699Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5251790Z 2025-12-04T09:45:06.5251839Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5252025Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5252272Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5252415Z 2025-12-04T09:45:06.5252490Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5252787Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5253011Z 2025-12-04T09:45:06.5253100Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5253305Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5253457Z unimplemented [] 2025-12-04T09:45:06.5253569Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5253705Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5253818Z graph_break [] 2025-12-04T09:45:06.5253980Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5254335Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5254645Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5254793Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5254944Z unimplemented [] 2025-12-04T09:45:06.5255058Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5255194Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5255302Z graph_break [] 2025-12-04T09:45:06.5255433Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5255777Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5256093Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5256241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5256397Z unimplemented [] 2025-12-04T09:45:06.5256514Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5256646Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5256788Z graph_break [] 2025-12-04T09:45:06.5256922Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5257261Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5257570Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5257718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5257870Z unimplemented [] 2025-12-04T09:45:06.5257984Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5258119Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5258273Z graph_break [] 2025-12-04T09:45:06.5258406Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5258750Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5259065Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5259213Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5259360Z unimplemented [] 2025-12-04T09:45:06.5259474Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5259608Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5259720Z graph_break [] 2025-12-04T09:45:06.5259854Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5260197Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5260511Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5260659Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5260809Z unimplemented [] 2025-12-04T09:45:06.5260924Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5261062Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5261171Z graph_break [] 2025-12-04T09:45:06.5261306Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5261650Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5261959Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5262106Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5262288Z unimplemented [] 2025-12-04T09:45:06.5262407Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5262544Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5262651Z graph_break [] 2025-12-04T09:45:06.5262784Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5263125Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5263433Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5263582Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5263734Z unimplemented [] 2025-12-04T09:45:06.5263850Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5263987Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5264098Z graph_break [] 2025-12-04T09:45:06.5264236Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5264588Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5264904Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5265058Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5265217Z unimplemented [] 2025-12-04T09:45:06.5265340Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5265483Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5265595Z graph_break [] 2025-12-04T09:45:06.5265732Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5266081Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5266396Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5266549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5266790Z unimplemented [] 2025-12-04T09:45:06.5266915Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5267061Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5267178Z graph_break [] 2025-12-04T09:45:06.5267316Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5267666Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5267981Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5268167Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5268366Z Traceback (most recent call last): 2025-12-04T09:45:06.5268572Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5268770Z return value(self) 2025-12-04T09:45:06.5268994Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5269234Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5269460Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5269698Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5269955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5270228Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5270391Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5270484Z 2025-12-04T09:45:06.5270531Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5270745Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5270991Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5271139Z 2025-12-04T09:45:06.5271214Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5271514Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5271739Z 2025-12-04T09:45:06.5271830Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5272034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5272188Z unimplemented [] 2025-12-04T09:45:06.5272308Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5272449Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5272563Z graph_break [] 2025-12-04T09:45:06.5272696Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5273049Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5273365Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5273518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5273673Z unimplemented [] 2025-12-04T09:45:06.5273793Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5273931Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5274043Z graph_break [] 2025-12-04T09:45:06.5274175Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5274516Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5274825Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5274977Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5275171Z unimplemented [] 2025-12-04T09:45:06.5275288Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5275428Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5275536Z graph_break [] 2025-12-04T09:45:06.5275666Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5276007Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5276316Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5276468Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5276621Z unimplemented [] 2025-12-04T09:45:06.5276773Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5276911Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5277024Z graph_break [] 2025-12-04T09:45:06.5277153Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5277501Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5277810Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5277962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5278114Z unimplemented [] 2025-12-04T09:45:06.5278231Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5278366Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5278476Z graph_break [] 2025-12-04T09:45:06.5278608Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5278982Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5279296Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5279447Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5279598Z unimplemented [] 2025-12-04T09:45:06.5279714Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5279851Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5279962Z graph_break [] 2025-12-04T09:45:06.5280091Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5280432Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5280740Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5280884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5281031Z unimplemented [] 2025-12-04T09:45:06.5281148Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5281285Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5281394Z graph_break [] 2025-12-04T09:45:06.5281525Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5281867Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5282176Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5282326Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5282476Z unimplemented [] 2025-12-04T09:45:06.5282592Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5282724Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5282833Z graph_break [] 2025-12-04T09:45:06.5282968Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5283310Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5283650Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5283797Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5283945Z unimplemented [] 2025-12-04T09:45:06.5284059Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5284193Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5284299Z graph_break [] 2025-12-04T09:45:06.5284425Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5284766Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5285069Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5285216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5285366Z unimplemented [] 2025-12-04T09:45:06.5285478Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5285612Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5285720Z graph_break [] 2025-12-04T09:45:06.5285847Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5286184Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5286489Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5286633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5286827Z unimplemented [] 2025-12-04T09:45:06.5286939Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5287073Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5287213Z graph_break [] 2025-12-04T09:45:06.5287342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5287686Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5287992Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5288169Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5288355Z Traceback (most recent call last): 2025-12-04T09:45:06.5288553Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5288746Z return value(self) 2025-12-04T09:45:06.5288958Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5289193Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5289421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5289657Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5289910Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5290175Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5290333Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5290421Z 2025-12-04T09:45:06.5290468Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5290645Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5290882Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5291019Z 2025-12-04T09:45:06.5291093Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5291388Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5291639Z 2025-12-04T09:45:06.5291728Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5291924Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5292070Z unimplemented [] 2025-12-04T09:45:06.5292181Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5292314Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5292420Z graph_break [] 2025-12-04T09:45:06.5292547Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5292885Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5293191Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5293336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5293489Z unimplemented [] 2025-12-04T09:45:06.5293599Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5293732Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5293839Z graph_break [] 2025-12-04T09:45:06.5293967Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5294303Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5294609Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5294753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5294900Z unimplemented [] 2025-12-04T09:45:06.5295011Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5295143Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5295248Z graph_break [] 2025-12-04T09:45:06.5295400Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5295743Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5296048Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5296192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5296338Z unimplemented [] 2025-12-04T09:45:06.5296450Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5296581Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5296686Z graph_break [] 2025-12-04T09:45:06.5296851Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5297191Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5297497Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5297645Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5297791Z unimplemented [] 2025-12-04T09:45:06.5297901Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5298033Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5298137Z graph_break [] 2025-12-04T09:45:06.5298262Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5298603Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5298906Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5299050Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5299195Z unimplemented [] 2025-12-04T09:45:06.5299310Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5299443Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5299586Z graph_break [] 2025-12-04T09:45:06.5299712Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5300049Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5300354Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5300497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5300644Z unimplemented [] 2025-12-04T09:45:06.5300755Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5300887Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5300992Z graph_break [] 2025-12-04T09:45:06.5301118Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5301456Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5301763Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5301906Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5302052Z unimplemented [] 2025-12-04T09:45:06.5302163Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5302295Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5302401Z graph_break [] 2025-12-04T09:45:06.5302527Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5302862Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5303165Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5303351Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5303498Z unimplemented [] 2025-12-04T09:45:06.5303614Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5303745Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5303850Z graph_break [] 2025-12-04T09:45:06.5303976Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5304312Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5304614Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5304757Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5304903Z unimplemented [] 2025-12-04T09:45:06.5305016Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5305147Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5305255Z graph_break [] 2025-12-04T09:45:06.5305387Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5305730Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5306034Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5306178Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5306326Z unimplemented [] 2025-12-04T09:45:06.5306438Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5306572Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5306678Z graph_break [] 2025-12-04T09:45:06.5306835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5307172Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5307479Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5307663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5307809Z unimplemented [] 2025-12-04T09:45:06.5307921Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5308056Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5308162Z graph_break [] 2025-12-04T09:45:06.5308288Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5308623Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5308925Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5309100Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5309285Z Traceback (most recent call last): 2025-12-04T09:45:06.5309483Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5309681Z return value(self) 2025-12-04T09:45:06.5309891Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5310124Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5310347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5310581Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5310835Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5311101Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5311261Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5311347Z 2025-12-04T09:45:06.5311394Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5311610Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5311852Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5311989Z 2025-12-04T09:45:06.5312064Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5312354Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5312571Z 2025-12-04T09:45:06.5312660Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5312856Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5313004Z unimplemented [] 2025-12-04T09:45:06.5313119Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5313253Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5313360Z graph_break [] 2025-12-04T09:45:06.5313491Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5313835Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5314140Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5314286Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5314432Z unimplemented [] 2025-12-04T09:45:06.5314545Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5314677Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5314713Z graph_break [] 2025-12-04T09:45:06.5314786Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5315019Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5315066Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5315139Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5315208Z unimplemented [] 2025-12-04T09:45:06.5315263Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5315307Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5315343Z graph_break [] 2025-12-04T09:45:06.5315416Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5315643Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5315687Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5315759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5315798Z unimplemented [] 2025-12-04T09:45:06.5315853Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5315898Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5315934Z graph_break [] 2025-12-04T09:45:06.5316011Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5316236Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5316280Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5316351Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5316390Z unimplemented [] 2025-12-04T09:45:06.5316445Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5316490Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5316525Z graph_break [] 2025-12-04T09:45:06.5316600Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5316889Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5316935Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5317009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5317047Z unimplemented [] 2025-12-04T09:45:06.5317102Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5317147Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5317183Z graph_break [] 2025-12-04T09:45:06.5317256Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5317481Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5317524Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5317596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5317637Z unimplemented [] 2025-12-04T09:45:06.5317692Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5317738Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5317775Z graph_break [] 2025-12-04T09:45:06.5317850Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5318075Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5318119Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5318191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5318230Z unimplemented [] 2025-12-04T09:45:06.5318284Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5318328Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5318364Z graph_break [] 2025-12-04T09:45:06.5318440Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5318665Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5318744Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5318817Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5318855Z unimplemented [] 2025-12-04T09:45:06.5318910Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5318954Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5318989Z graph_break [] 2025-12-04T09:45:06.5319064Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5319291Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5319334Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5319408Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5319448Z unimplemented [] 2025-12-04T09:45:06.5319502Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5319545Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5319581Z graph_break [] 2025-12-04T09:45:06.5319654Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5319879Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5319923Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5319995Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5320034Z unimplemented [] 2025-12-04T09:45:06.5320089Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5320133Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5320189Z graph_break [] 2025-12-04T09:45:06.5320268Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5320494Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5320543Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5320616Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5320658Z unimplemented [] 2025-12-04T09:45:06.5320714Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5320761Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5320798Z graph_break [] 2025-12-04T09:45:06.5320875Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5321104Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5321155Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5321230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5321271Z unimplemented [] 2025-12-04T09:45:06.5321327Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5321374Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5321413Z graph_break [] 2025-12-04T09:45:06.5321492Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5321718Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5321764Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5321870Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5321920Z Traceback (most recent call last): 2025-12-04T09:45:06.5322040Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5322109Z return value(self) 2025-12-04T09:45:06.5322258Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5322314Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5322455Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5322516Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5322679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5322754Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5322811Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5322813Z 2025-12-04T09:45:06.5322860Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5322970Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5323073Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5323075Z 2025-12-04T09:45:06.5323152Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5323339Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5323341Z 2025-12-04T09:45:06.5323434Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5323509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5323552Z unimplemented [] 2025-12-04T09:45:06.5323609Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5323656Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5323727Z graph_break [] 2025-12-04T09:45:06.5323805Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5324039Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5324087Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5324162Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5324205Z unimplemented [] 2025-12-04T09:45:06.5324264Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5324311Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5324349Z graph_break [] 2025-12-04T09:45:06.5324427Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5324658Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5324708Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5324781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5324823Z unimplemented [] 2025-12-04T09:45:06.5324881Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5324929Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5324968Z graph_break [] 2025-12-04T09:45:06.5325045Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5325274Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5325320Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5325394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5325438Z unimplemented [] 2025-12-04T09:45:06.5325497Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5325571Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5325611Z graph_break [] 2025-12-04T09:45:06.5325688Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5325915Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5325960Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5326033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5326076Z unimplemented [] 2025-12-04T09:45:06.5326132Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5326181Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5326220Z graph_break [] 2025-12-04T09:45:06.5326298Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5326525Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5326574Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5326648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5326690Z unimplemented [] 2025-12-04T09:45:06.5326782Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5326829Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5326867Z graph_break [] 2025-12-04T09:45:06.5326945Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5327173Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5327221Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5327324Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5327367Z unimplemented [] 2025-12-04T09:45:06.5327424Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5327470Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5327506Z graph_break [] 2025-12-04T09:45:06.5327585Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5327816Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5327861Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5327939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5327978Z unimplemented [] 2025-12-04T09:45:06.5328038Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5328083Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5328121Z graph_break [] 2025-12-04T09:45:06.5328199Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5328431Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5328475Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5328553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5328591Z unimplemented [] 2025-12-04T09:45:06.5328652Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5328695Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5328734Z graph_break [] 2025-12-04T09:45:06.5328808Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5329042Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5329086Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5329198Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5329237Z unimplemented [] 2025-12-04T09:45:06.5329297Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5329341Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5329382Z graph_break [] 2025-12-04T09:45:06.5329457Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5329689Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5329733Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5329811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5329850Z unimplemented [] 2025-12-04T09:45:06.5329912Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5329960Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5330003Z graph_break [] 2025-12-04T09:45:06.5330077Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5330308Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5330352Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5330431Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5330470Z unimplemented [] 2025-12-04T09:45:06.5330530Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5330574Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5330614Z graph_break [] 2025-12-04T09:45:06.5330690Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5330942Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5330988Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5331064Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5331102Z unimplemented [] 2025-12-04T09:45:06.5331162Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5331205Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5331247Z graph_break [] 2025-12-04T09:45:06.5331321Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5331552Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5331595Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5331674Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5331714Z unimplemented [] 2025-12-04T09:45:06.5331775Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5331818Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5331859Z graph_break [] 2025-12-04T09:45:06.5331934Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5332167Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5332212Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5332322Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5332369Z Traceback (most recent call last): 2025-12-04T09:45:06.5332494Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5332534Z return value(self) 2025-12-04T09:45:06.5332689Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5332774Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5332916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5332978Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5333141Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5333214Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5333272Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5333274Z 2025-12-04T09:45:06.5333319Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5333425Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5333529Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5333533Z 2025-12-04T09:45:06.5333612Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5333798Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5333800Z 2025-12-04T09:45:06.5333887Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5333967Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5334007Z unimplemented [] 2025-12-04T09:45:06.5334068Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5334113Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5334154Z graph_break [] 2025-12-04T09:45:06.5334228Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5334481Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5334528Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5334602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5334641Z unimplemented [] 2025-12-04T09:45:06.5334698Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5334740Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5334780Z graph_break [] 2025-12-04T09:45:06.5334853Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5335083Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5335125Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5335202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5335240Z unimplemented [] 2025-12-04T09:45:06.5335299Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5335341Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5335379Z graph_break [] 2025-12-04T09:45:06.5335452Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5335683Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5335725Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5335800Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5335837Z unimplemented [] 2025-12-04T09:45:06.5335894Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5335936Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5335973Z graph_break [] 2025-12-04T09:45:06.5336048Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5336311Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5336353Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5336428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5336465Z unimplemented [] 2025-12-04T09:45:06.5336521Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5336563Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5336602Z graph_break [] 2025-12-04T09:45:06.5336674Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5336935Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5336979Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5337056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5337093Z unimplemented [] 2025-12-04T09:45:06.5337149Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5337192Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5337229Z graph_break [] 2025-12-04T09:45:06.5337301Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5337528Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5337569Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5337644Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5337681Z unimplemented [] 2025-12-04T09:45:06.5337737Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5337808Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5337849Z graph_break [] 2025-12-04T09:45:06.5337922Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5338150Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5338191Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5338266Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5338303Z unimplemented [] 2025-12-04T09:45:06.5338360Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5338403Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5338441Z graph_break [] 2025-12-04T09:45:06.5338513Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5338745Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5338790Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5338864Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5338901Z unimplemented [] 2025-12-04T09:45:06.5338957Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5339000Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5339038Z graph_break [] 2025-12-04T09:45:06.5339111Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5339338Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5339379Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5339456Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5339497Z unimplemented [] 2025-12-04T09:45:06.5339585Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5339627Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5339665Z graph_break [] 2025-12-04T09:45:06.5339738Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5339966Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5340007Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5340082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5340119Z unimplemented [] 2025-12-04T09:45:06.5340176Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5340218Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5340256Z graph_break [] 2025-12-04T09:45:06.5340330Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5340561Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5340602Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5340676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5340713Z unimplemented [] 2025-12-04T09:45:06.5340770Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5340812Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5340851Z graph_break [] 2025-12-04T09:45:06.5340923Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5341151Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5341217Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5341292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5341333Z unimplemented [] 2025-12-04T09:45:06.5341388Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5341431Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5341468Z graph_break [] 2025-12-04T09:45:06.5341541Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5341768Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5341812Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5341885Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5341924Z unimplemented [] 2025-12-04T09:45:06.5341979Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5342024Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5342061Z graph_break [] 2025-12-04T09:45:06.5342137Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5342363Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5342406Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5342478Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5342517Z unimplemented [] 2025-12-04T09:45:06.5342572Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5342616Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5342652Z graph_break [] 2025-12-04T09:45:06.5342726Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5342954Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5343024Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5343128Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5343175Z Traceback (most recent call last): 2025-12-04T09:45:06.5343293Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5343334Z return value(self) 2025-12-04T09:45:06.5343481Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5343535Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5343672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5343733Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5343891Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5343966Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5344018Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5344020Z 2025-12-04T09:45:06.5344065Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5344167Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5344270Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5344272Z 2025-12-04T09:45:06.5344345Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5344531Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5344534Z 2025-12-04T09:45:06.5344620Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5344716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5344756Z unimplemented [] 2025-12-04T09:45:06.5344814Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5344857Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5344895Z graph_break [] 2025-12-04T09:45:06.5344968Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5345198Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5345241Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5345316Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5345353Z unimplemented [] 2025-12-04T09:45:06.5345410Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5345453Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5345493Z graph_break [] 2025-12-04T09:45:06.5345565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5345797Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5345839Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5345913Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5345950Z unimplemented [] 2025-12-04T09:45:06.5346007Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5346050Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5346088Z graph_break [] 2025-12-04T09:45:06.5346160Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5346390Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5346460Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5346534Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5346573Z unimplemented [] 2025-12-04T09:45:06.5346628Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5346672Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5346708Z graph_break [] 2025-12-04T09:45:06.5346821Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5347045Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5347089Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5347161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5347200Z unimplemented [] 2025-12-04T09:45:06.5347256Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5347302Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5347338Z graph_break [] 2025-12-04T09:45:06.5347411Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5347636Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5347679Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5347751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5347790Z unimplemented [] 2025-12-04T09:45:06.5347845Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5347889Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5347924Z graph_break [] 2025-12-04T09:45:06.5347998Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5348255Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5348303Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5348375Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5348414Z unimplemented [] 2025-12-04T09:45:06.5348469Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5348515Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5348551Z graph_break [] 2025-12-04T09:45:06.5348626Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5348853Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5348896Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5348969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5349010Z unimplemented [] 2025-12-04T09:45:06.5349065Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5349109Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5349145Z graph_break [] 2025-12-04T09:45:06.5349219Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5349444Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5349488Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5349560Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5349599Z unimplemented [] 2025-12-04T09:45:06.5349654Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5349698Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5349735Z graph_break [] 2025-12-04T09:45:06.5349809Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5350075Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5350119Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5350192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5350230Z unimplemented [] 2025-12-04T09:45:06.5350285Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5350331Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5350367Z graph_break [] 2025-12-04T09:45:06.5350442Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5350670Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5350715Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5350788Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5350826Z unimplemented [] 2025-12-04T09:45:06.5350881Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5350926Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5350961Z graph_break [] 2025-12-04T09:45:06.5351035Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5351261Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5351306Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5351377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5351416Z unimplemented [] 2025-12-04T09:45:06.5351494Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5351540Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5351576Z graph_break [] 2025-12-04T09:45:06.5351650Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5351875Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5351918Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5351990Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5352029Z unimplemented [] 2025-12-04T09:45:06.5352084Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5352130Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5352165Z graph_break [] 2025-12-04T09:45:06.5352240Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5352468Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5352514Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5352587Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5352626Z unimplemented [] 2025-12-04T09:45:06.5352680Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5352725Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5352761Z graph_break [] 2025-12-04T09:45:06.5352834Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5353060Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5353105Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5353178Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5353247Z unimplemented [] 2025-12-04T09:45:06.5353302Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5353347Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5353383Z graph_break [] 2025-12-04T09:45:06.5353458Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5353684Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5353728Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5353801Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5353840Z unimplemented [] 2025-12-04T09:45:06.5353894Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5353939Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5353976Z graph_break [] 2025-12-04T09:45:06.5354051Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5354279Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5354323Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5354427Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5354474Z Traceback (most recent call last): 2025-12-04T09:45:06.5354595Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5354633Z return value(self) 2025-12-04T09:45:06.5354785Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5354837Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5354998Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5355060Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5355220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5355293Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5355348Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5355350Z 2025-12-04T09:45:06.5355394Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5355498Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5355599Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5355601Z 2025-12-04T09:45:06.5355676Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5355860Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5355863Z 2025-12-04T09:45:06.5355951Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5356024Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5356063Z unimplemented [] 2025-12-04T09:45:06.5356119Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5356164Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5356200Z graph_break [] 2025-12-04T09:45:06.5356276Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5356503Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5356548Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5356623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5356662Z unimplemented [] 2025-12-04T09:45:06.5356778Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5356824Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5356861Z graph_break [] 2025-12-04T09:45:06.5356937Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5357166Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5357209Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5357281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5357322Z unimplemented [] 2025-12-04T09:45:06.5357377Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5357424Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5357461Z graph_break [] 2025-12-04T09:45:06.5357537Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5357764Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5357809Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5357881Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5357921Z unimplemented [] 2025-12-04T09:45:06.5357976Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5358020Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5358055Z graph_break [] 2025-12-04T09:45:06.5358130Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5358355Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5358429Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5358504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5358543Z unimplemented [] 2025-12-04T09:45:06.5358598Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5358643Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5358678Z graph_break [] 2025-12-04T09:45:06.5358753Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5358981Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5359026Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5359098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5359137Z unimplemented [] 2025-12-04T09:45:06.5359193Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5359239Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5359276Z graph_break [] 2025-12-04T09:45:06.5359351Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5359577Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5359620Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5359694Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5359731Z unimplemented [] 2025-12-04T09:45:06.5359787Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5359830Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5359865Z graph_break [] 2025-12-04T09:45:06.5359939Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5360168Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5360242Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5360317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5360355Z unimplemented [] 2025-12-04T09:45:06.5360411Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5360453Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5360490Z graph_break [] 2025-12-04T09:45:06.5360565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5371579Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5371646Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5371736Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5371777Z unimplemented [] 2025-12-04T09:45:06.5371847Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5371896Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5371960Z graph_break [] 2025-12-04T09:45:06.5372039Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5372283Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5372331Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5372410Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5372453Z unimplemented [] 2025-12-04T09:45:06.5372514Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5372558Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5372597Z graph_break [] 2025-12-04T09:45:06.5372726Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5372962Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5373005Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5373080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5373117Z unimplemented [] 2025-12-04T09:45:06.5373176Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5373219Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5373258Z graph_break [] 2025-12-04T09:45:06.5373330Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5373562Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5373607Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5373683Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5373720Z unimplemented [] 2025-12-04T09:45:06.5373777Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5373820Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5373858Z graph_break [] 2025-12-04T09:45:06.5373931Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5374160Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5374202Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5374277Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5374314Z unimplemented [] 2025-12-04T09:45:06.5374369Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5374414Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5374489Z graph_break [] 2025-12-04T09:45:06.5374562Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5374789Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5374830Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5374904Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5374941Z unimplemented [] 2025-12-04T09:45:06.5374997Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5375039Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5375076Z graph_break [] 2025-12-04T09:45:06.5375147Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5375376Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5375421Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5375495Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5375531Z unimplemented [] 2025-12-04T09:45:06.5375589Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5375631Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5375669Z graph_break [] 2025-12-04T09:45:06.5375742Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5375969Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5376010Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5376084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5376178Z unimplemented [] 2025-12-04T09:45:06.5376236Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5376278Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5376315Z graph_break [] 2025-12-04T09:45:06.5376387Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5376612Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5376653Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5376728Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5376803Z unimplemented [] 2025-12-04T09:45:06.5376860Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5376901Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5376939Z graph_break [] 2025-12-04T09:45:06.5377015Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5377244Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5377285Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5377392Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5377440Z Traceback (most recent call last): 2025-12-04T09:45:06.5377567Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5377606Z return value(self) 2025-12-04T09:45:06.5377760Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5377813Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5377958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5378022Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5378222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5378296Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5378353Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5378356Z 2025-12-04T09:45:06.5378401Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5378506Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5378609Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5378612Z 2025-12-04T09:45:06.5378686Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5378873Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5378876Z 2025-12-04T09:45:06.5378966Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5379041Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5379080Z unimplemented [] 2025-12-04T09:45:06.5379138Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5379181Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5379219Z graph_break [] 2025-12-04T09:45:06.5379291Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5379523Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5379566Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5379639Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5379677Z unimplemented [] 2025-12-04T09:45:06.5379766Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5379813Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5379851Z graph_break [] 2025-12-04T09:45:06.5379923Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5380153Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5380196Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5380270Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5380307Z unimplemented [] 2025-12-04T09:45:06.5380365Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5380408Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5380445Z graph_break [] 2025-12-04T09:45:06.5380517Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5380745Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5380789Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5380863Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5380900Z unimplemented [] 2025-12-04T09:45:06.5380956Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5380999Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5381037Z graph_break [] 2025-12-04T09:45:06.5381108Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5381336Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5381377Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5381454Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5381521Z unimplemented [] 2025-12-04T09:45:06.5381578Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5381620Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5381660Z graph_break [] 2025-12-04T09:45:06.5381732Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5381961Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5382002Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5382077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5382114Z unimplemented [] 2025-12-04T09:45:06.5382172Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5382216Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5382257Z graph_break [] 2025-12-04T09:45:06.5382329Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5382560Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5382601Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5382674Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5382710Z unimplemented [] 2025-12-04T09:45:06.5382766Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5382808Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5382845Z graph_break [] 2025-12-04T09:45:06.5382917Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5383165Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5383208Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5383281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5383317Z unimplemented [] 2025-12-04T09:45:06.5383374Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5383416Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5383453Z graph_break [] 2025-12-04T09:45:06.5383527Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5383756Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5383798Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5383871Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5383909Z unimplemented [] 2025-12-04T09:45:06.5383967Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5384012Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5384050Z graph_break [] 2025-12-04T09:45:06.5384124Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5384352Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5384395Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5384467Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5384507Z unimplemented [] 2025-12-04T09:45:06.5384561Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5384603Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5384640Z graph_break [] 2025-12-04T09:45:06.5384712Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5384940Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5385012Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5385085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5385123Z unimplemented [] 2025-12-04T09:45:06.5385178Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5385222Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5385257Z graph_break [] 2025-12-04T09:45:06.5385333Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5385560Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5385603Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5385678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5385721Z unimplemented [] 2025-12-04T09:45:06.5385777Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5385821Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5385857Z graph_break [] 2025-12-04T09:45:06.5385931Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5386157Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5386201Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5386276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5386316Z unimplemented [] 2025-12-04T09:45:06.5386371Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5386416Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5386483Z graph_break [] 2025-12-04T09:45:06.5386559Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5386832Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5386877Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5386950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5386990Z unimplemented [] 2025-12-04T09:45:06.5387046Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5387094Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5387132Z graph_break [] 2025-12-04T09:45:06.5387209Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5387440Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5387488Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5387561Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5387603Z unimplemented [] 2025-12-04T09:45:06.5387659Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5387705Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5387742Z graph_break [] 2025-12-04T09:45:06.5387817Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5388045Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5388091Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5388163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5388205Z unimplemented [] 2025-12-04T09:45:06.5388266Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5388347Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5388385Z graph_break [] 2025-12-04T09:45:06.5388461Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5388689Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5388735Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5388808Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5388848Z unimplemented [] 2025-12-04T09:45:06.5388903Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5388949Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5388985Z graph_break [] 2025-12-04T09:45:06.5389061Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5389288Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5389336Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5389409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5389449Z unimplemented [] 2025-12-04T09:45:06.5389504Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5389550Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5389587Z graph_break [] 2025-12-04T09:45:06.5389663Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5389890Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5389937Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5390039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5390081Z unimplemented [] 2025-12-04T09:45:06.5390137Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5390183Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5390220Z graph_break [] 2025-12-04T09:45:06.5390296Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5390523Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5390569Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5390673Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5390724Z Traceback (most recent call last): 2025-12-04T09:45:06.5390846Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5390892Z return value(self) 2025-12-04T09:45:06.5391044Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5391104Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5391245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5391314Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5391476Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5391558Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5391613Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5391620Z 2025-12-04T09:45:06.5391666Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5391775Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5391880Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5391909Z 2025-12-04T09:45:06.5391988Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5392173Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5392175Z 2025-12-04T09:45:06.5392268Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5392346Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5392392Z unimplemented [] 2025-12-04T09:45:06.5392451Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5392500Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5392540Z graph_break [] 2025-12-04T09:45:06.5392620Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5392853Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5392904Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5392980Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5393027Z unimplemented [] 2025-12-04T09:45:06.5393085Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5393135Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5393173Z graph_break [] 2025-12-04T09:45:06.5393254Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5393484Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5393536Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5393635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5393681Z unimplemented [] 2025-12-04T09:45:06.5393741Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5393791Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5393830Z graph_break [] 2025-12-04T09:45:06.5393910Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5394137Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5394187Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5394260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5394308Z unimplemented [] 2025-12-04T09:45:06.5394365Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5394414Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5394453Z graph_break [] 2025-12-04T09:45:06.5394528Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5394760Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5394803Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5394880Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5394918Z unimplemented [] 2025-12-04T09:45:06.5394978Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5395023Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5395064Z graph_break [] 2025-12-04T09:45:06.5395138Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5395368Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5395414Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5395517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5395556Z unimplemented [] 2025-12-04T09:45:06.5395616Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5395660Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5395700Z graph_break [] 2025-12-04T09:45:06.5395775Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5396006Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5396048Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5396123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5396162Z unimplemented [] 2025-12-04T09:45:06.5396222Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5396269Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5396314Z graph_break [] 2025-12-04T09:45:06.5396388Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5396619Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5396662Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5396771Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5396810Z unimplemented [] 2025-12-04T09:45:06.5396872Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5396916Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5396957Z graph_break [] 2025-12-04T09:45:06.5397031Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5397295Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5397341Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5397419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5397459Z unimplemented [] 2025-12-04T09:45:06.5397520Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5397565Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5397605Z graph_break [] 2025-12-04T09:45:06.5397680Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5397910Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5397953Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5398034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5398072Z unimplemented [] 2025-12-04T09:45:06.5398137Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5398182Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5398224Z graph_break [] 2025-12-04T09:45:06.5398302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5398529Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5398572Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5398647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5398685Z unimplemented [] 2025-12-04T09:45:06.5398741Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5398784Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5398821Z graph_break [] 2025-12-04T09:45:06.5398896Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5399164Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5399205Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5399279Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5399317Z unimplemented [] 2025-12-04T09:45:06.5399374Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5399417Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5399455Z graph_break [] 2025-12-04T09:45:06.5399528Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5399755Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5399798Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5399874Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5399910Z unimplemented [] 2025-12-04T09:45:06.5399967Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5400009Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5400046Z graph_break [] 2025-12-04T09:45:06.5400118Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5400346Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5400388Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5400462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5400499Z unimplemented [] 2025-12-04T09:45:06.5400556Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5400627Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5400667Z graph_break [] 2025-12-04T09:45:06.5400739Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5400968Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5401010Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5401083Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5401120Z unimplemented [] 2025-12-04T09:45:06.5401177Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5401219Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5401256Z graph_break [] 2025-12-04T09:45:06.5401327Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5401557Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5401600Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5401673Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5401710Z unimplemented [] 2025-12-04T09:45:06.5401766Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5401808Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5401846Z graph_break [] 2025-12-04T09:45:06.5401917Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5402145Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5402186Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5402260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5402299Z unimplemented [] 2025-12-04T09:45:06.5402386Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5402429Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5402466Z graph_break [] 2025-12-04T09:45:06.5402539Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5402769Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5402811Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5402884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5402922Z unimplemented [] 2025-12-04T09:45:06.5402979Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5403021Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5403060Z graph_break [] 2025-12-04T09:45:06.5403134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5403362Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5403404Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5403478Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5403515Z unimplemented [] 2025-12-04T09:45:06.5403572Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5403615Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5403652Z graph_break [] 2025-12-04T09:45:06.5403725Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5403955Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5404017Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5404091Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5404130Z unimplemented [] 2025-12-04T09:45:06.5404187Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5404229Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5404267Z graph_break [] 2025-12-04T09:45:06.5404339Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5404568Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5404610Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5404716Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5404761Z Traceback (most recent call last): 2025-12-04T09:45:06.5404883Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5404921Z return value(self) 2025-12-04T09:45:06.5405073Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5405125Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5405263Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5405323Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5405483Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5405556Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5405608Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5405610Z 2025-12-04T09:45:06.5405654Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5405757Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5405860Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5405890Z 2025-12-04T09:45:06.5405963Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5406147Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5406149Z 2025-12-04T09:45:06.5406235Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5406310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5406348Z unimplemented [] 2025-12-04T09:45:06.5406407Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5406450Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5406488Z graph_break [] 2025-12-04T09:45:06.5406560Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5406922Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5406968Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5407042Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5407079Z unimplemented [] 2025-12-04T09:45:06.5407138Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5407181Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5407219Z graph_break [] 2025-12-04T09:45:06.5407290Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5407519Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5407562Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5407667Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5407707Z unimplemented [] 2025-12-04T09:45:06.5407765Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5407808Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5407846Z graph_break [] 2025-12-04T09:45:06.5407918Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5408148Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5408190Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5408263Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5408300Z unimplemented [] 2025-12-04T09:45:06.5408356Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5408399Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5408440Z graph_break [] 2025-12-04T09:45:06.5408512Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5408742Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5408783Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5408856Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5408893Z unimplemented [] 2025-12-04T09:45:06.5408949Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5408992Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5409029Z graph_break [] 2025-12-04T09:45:06.5409101Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5409331Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5409407Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5409481Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5409519Z unimplemented [] 2025-12-04T09:45:06.5409575Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5409617Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5409655Z graph_break [] 2025-12-04T09:45:06.5409727Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5409955Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5409997Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5410071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5410108Z unimplemented [] 2025-12-04T09:45:06.5410166Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5410211Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5410248Z graph_break [] 2025-12-04T09:45:06.5410321Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5410549Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5410591Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5410665Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5410702Z unimplemented [] 2025-12-04T09:45:06.5410759Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5410801Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5410839Z graph_break [] 2025-12-04T09:45:06.5410911Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5411161Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5411207Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5411279Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5411317Z unimplemented [] 2025-12-04T09:45:06.5411372Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5411414Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5411452Z graph_break [] 2025-12-04T09:45:06.5411523Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5411751Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5411795Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5411868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5411909Z unimplemented [] 2025-12-04T09:45:06.5411964Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5412008Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5412044Z graph_break [] 2025-12-04T09:45:06.5412119Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5412345Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5412388Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5412461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5412499Z unimplemented [] 2025-12-04T09:45:06.5412553Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5412598Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5412635Z graph_break [] 2025-12-04T09:45:06.5412710Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5412969Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5413014Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5413086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5413125Z unimplemented [] 2025-12-04T09:45:06.5413179Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5413224Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5413260Z graph_break [] 2025-12-04T09:45:06.5413333Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5413563Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5413607Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5413681Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5413719Z unimplemented [] 2025-12-04T09:45:06.5413774Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5413817Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5413853Z graph_break [] 2025-12-04T09:45:06.5413928Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5414153Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5414197Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5414269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5414307Z unimplemented [] 2025-12-04T09:45:06.5414384Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5414432Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5414467Z graph_break [] 2025-12-04T09:45:06.5414540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5414765Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5414808Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5414880Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5414918Z unimplemented [] 2025-12-04T09:45:06.5414972Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5415016Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5415052Z graph_break [] 2025-12-04T09:45:06.5415125Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5415352Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5415396Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5415468Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5415507Z unimplemented [] 2025-12-04T09:45:06.5415562Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5415606Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5415641Z graph_break [] 2025-12-04T09:45:06.5415714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5415939Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5415983Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5416056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5416121Z unimplemented [] 2025-12-04T09:45:06.5416177Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5416220Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5416256Z graph_break [] 2025-12-04T09:45:06.5416330Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5416557Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5416601Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5416673Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5416713Z unimplemented [] 2025-12-04T09:45:06.5416804Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5416849Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5416884Z graph_break [] 2025-12-04T09:45:06.5416960Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5417186Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5417229Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5417301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5417339Z unimplemented [] 2025-12-04T09:45:06.5417394Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5417438Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5417473Z graph_break [] 2025-12-04T09:45:06.5417547Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5417802Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5417846Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5417921Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5417959Z unimplemented [] 2025-12-04T09:45:06.5418013Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5418059Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5418095Z graph_break [] 2025-12-04T09:45:06.5418169Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5418395Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5418438Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5418510Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5418549Z unimplemented [] 2025-12-04T09:45:06.5418605Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5418649Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5418686Z graph_break [] 2025-12-04T09:45:06.5418759Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5418984Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5419027Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5419132Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5419179Z Traceback (most recent call last): 2025-12-04T09:45:06.5419297Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5419337Z return value(self) 2025-12-04T09:45:06.5419485Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5419539Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5419711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5419771Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5419930Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5420002Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5420055Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5420057Z 2025-12-04T09:45:06.5420101Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5420205Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5420306Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5420308Z 2025-12-04T09:45:06.5420383Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5420568Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5420570Z 2025-12-04T09:45:06.5420657Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5420730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5420771Z unimplemented [] 2025-12-04T09:45:06.5420826Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5420870Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5420906Z graph_break [] 2025-12-04T09:45:06.5420979Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5421209Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5421277Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5421351Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5421390Z unimplemented [] 2025-12-04T09:45:06.5421445Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5421488Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5421524Z graph_break [] 2025-12-04T09:45:06.5421597Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5421824Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5421869Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5421941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5421980Z unimplemented [] 2025-12-04T09:45:06.5422034Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5422079Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5422116Z graph_break [] 2025-12-04T09:45:06.5422189Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5422418Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5422462Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5422536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5422573Z unimplemented [] 2025-12-04T09:45:06.5422628Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5422671Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5422706Z graph_break [] 2025-12-04T09:45:06.5422779Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5423007Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5423077Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5423149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5423188Z unimplemented [] 2025-12-04T09:45:06.5423243Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5423290Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5423325Z graph_break [] 2025-12-04T09:45:06.5423398Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5423622Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5423666Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5423739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5423778Z unimplemented [] 2025-12-04T09:45:06.5423835Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5423879Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5423914Z graph_break [] 2025-12-04T09:45:06.5423989Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5424215Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5424259Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5424331Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5424370Z unimplemented [] 2025-12-04T09:45:06.5424424Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5424468Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5424503Z graph_break [] 2025-12-04T09:45:06.5424598Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5424829Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5424871Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5424945Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5424982Z unimplemented [] 2025-12-04T09:45:06.5425038Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5425080Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5425116Z graph_break [] 2025-12-04T09:45:06.5425188Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5425416Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5425459Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5425534Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5425571Z unimplemented [] 2025-12-04T09:45:06.5425628Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5425670Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5425707Z graph_break [] 2025-12-04T09:45:06.5425780Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5426007Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5426048Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5426124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5426161Z unimplemented [] 2025-12-04T09:45:06.5426217Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5426261Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5426332Z graph_break [] 2025-12-04T09:45:06.5426404Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5426632Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5426673Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5426790Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5426827Z unimplemented [] 2025-12-04T09:45:06.5426884Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5426926Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5426964Z graph_break [] 2025-12-04T09:45:06.5427036Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5427264Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5427307Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5427381Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5427418Z unimplemented [] 2025-12-04T09:45:06.5427474Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5427516Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5427554Z graph_break [] 2025-12-04T09:45:06.5427626Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5427943Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5428263Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5428414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5428599Z unimplemented [] 2025-12-04T09:45:06.5428716Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5428854Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5428963Z graph_break [] 2025-12-04T09:45:06.5429091Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5429432Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5429743Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5429887Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5430036Z unimplemented [] 2025-12-04T09:45:06.5430150Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5430285Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5430392Z graph_break [] 2025-12-04T09:45:06.5430523Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5430864Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5431170Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5431317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5431465Z unimplemented [] 2025-12-04T09:45:06.5431579Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5431713Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5431823Z graph_break [] 2025-12-04T09:45:06.5431952Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5432293Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5432603Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5432748Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5432937Z unimplemented [] 2025-12-04T09:45:06.5433049Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5433181Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5433288Z graph_break [] 2025-12-04T09:45:06.5433415Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5433752Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5434057Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5434201Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5434350Z unimplemented [] 2025-12-04T09:45:06.5434461Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5434595Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5434702Z graph_break [] 2025-12-04T09:45:06.5434837Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5435176Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5435481Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5435625Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5435772Z unimplemented [] 2025-12-04T09:45:06.5435882Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5436015Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5436120Z graph_break [] 2025-12-04T09:45:06.5436246Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5436625Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5436977Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5437126Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5437279Z unimplemented [] 2025-12-04T09:45:06.5437397Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5437535Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5437647Z graph_break [] 2025-12-04T09:45:06.5437780Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5438126Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5438437Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5438585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5438740Z unimplemented [] 2025-12-04T09:45:06.5438854Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5438995Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5439105Z graph_break [] 2025-12-04T09:45:06.5439239Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5439588Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5439897Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5440046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5440198Z unimplemented [] 2025-12-04T09:45:06.5440316Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5440455Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5440567Z graph_break [] 2025-12-04T09:45:06.5440699Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5441043Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5441406Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5441554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5441705Z unimplemented [] 2025-12-04T09:45:06.5441821Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5441957Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5442069Z graph_break [] 2025-12-04T09:45:06.5442199Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5442541Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5442855Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5443035Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5443233Z Traceback (most recent call last): 2025-12-04T09:45:06.5443434Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5443628Z return value(self) 2025-12-04T09:45:06.5443844Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5444083Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5444314Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5444552Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5444813Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5445114Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5445278Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5445376Z 2025-12-04T09:45:06.5445423Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5445607Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5445852Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5445992Z 2025-12-04T09:45:06.5446070Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5446368Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5446590Z 2025-12-04T09:45:06.5446681Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5446926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5447084Z unimplemented [] 2025-12-04T09:45:06.5447207Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5447352Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5447464Z graph_break [] 2025-12-04T09:45:06.5447597Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5447944Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5448416Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5448567Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5448719Z unimplemented [] 2025-12-04T09:45:06.5448838Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5448978Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5449089Z graph_break [] 2025-12-04T09:45:06.5449223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5449568Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5449916Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5450065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5450217Z unimplemented [] 2025-12-04T09:45:06.5450335Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5450475Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5450587Z graph_break [] 2025-12-04T09:45:06.5450725Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5451074Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5451385Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5451535Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5451690Z unimplemented [] 2025-12-04T09:45:06.5451807Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5451947Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5452057Z graph_break [] 2025-12-04T09:45:06.5452191Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5452535Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5452845Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5452994Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5453146Z unimplemented [] 2025-12-04T09:45:06.5453258Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5453401Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5453557Z graph_break [] 2025-12-04T09:45:06.5453689Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5454039Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5454348Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5454495Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5454646Z unimplemented [] 2025-12-04T09:45:06.5454763Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5454902Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5455014Z graph_break [] 2025-12-04T09:45:06.5455147Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5455494Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5455805Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5455952Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5456106Z unimplemented [] 2025-12-04T09:45:06.5456222Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5456360Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5456472Z graph_break [] 2025-12-04T09:45:06.5456604Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5456969Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5457279Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5457429Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5457583Z unimplemented [] 2025-12-04T09:45:06.5457702Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5457880Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5457988Z graph_break [] 2025-12-04T09:45:06.5458122Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5458464Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5458775Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5458924Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5459076Z unimplemented [] 2025-12-04T09:45:06.5459192Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5459328Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5459438Z graph_break [] 2025-12-04T09:45:06.5459568Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5459912Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5460222Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5460371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5460520Z unimplemented [] 2025-12-04T09:45:06.5460634Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5460767Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5460877Z graph_break [] 2025-12-04T09:45:06.5461008Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5461354Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5461665Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5461847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5462002Z unimplemented [] 2025-12-04T09:45:06.5462118Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5462257Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5462368Z graph_break [] 2025-12-04T09:45:06.5462501Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5462847Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5463152Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5463298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5463445Z unimplemented [] 2025-12-04T09:45:06.5463557Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5463690Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5463797Z graph_break [] 2025-12-04T09:45:06.5463927Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5464272Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5464577Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5464721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5464868Z unimplemented [] 2025-12-04T09:45:06.5464982Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5465116Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5465226Z graph_break [] 2025-12-04T09:45:06.5465354Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5465693Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5465998Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5466175Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5466324Z unimplemented [] 2025-12-04T09:45:06.5466434Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5466566Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5466673Z graph_break [] 2025-12-04T09:45:06.5466832Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5467169Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5467475Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5467621Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5467768Z unimplemented [] 2025-12-04T09:45:06.5467885Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5468019Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5468130Z graph_break [] 2025-12-04T09:45:06.5468257Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5468595Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5468899Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5469043Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5469191Z unimplemented [] 2025-12-04T09:45:06.5469301Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5469433Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5469538Z graph_break [] 2025-12-04T09:45:06.5469663Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5470038Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5470345Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5470490Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5470635Z unimplemented [] 2025-12-04T09:45:06.5470749Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5470881Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5470987Z graph_break [] 2025-12-04T09:45:06.5471112Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5471448Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5471752Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5471901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5472050Z unimplemented [] 2025-12-04T09:45:06.5472161Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5472293Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5472398Z graph_break [] 2025-12-04T09:45:06.5472522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5472861Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5473165Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5473308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5473456Z unimplemented [] 2025-12-04T09:45:06.5473567Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5473699Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5473807Z graph_break [] 2025-12-04T09:45:06.5473935Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5474310Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5474613Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5474758Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5474905Z unimplemented [] 2025-12-04T09:45:06.5475017Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5475149Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5475255Z graph_break [] 2025-12-04T09:45:06.5475380Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5475720Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5476025Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5476173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5476320Z unimplemented [] 2025-12-04T09:45:06.5476431Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5476564Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5476669Z graph_break [] 2025-12-04T09:45:06.5476831Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5477169Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5477475Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5477619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5477766Z unimplemented [] 2025-12-04T09:45:06.5477910Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5478043Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5478154Z graph_break [] 2025-12-04T09:45:06.5478279Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5478617Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5478920Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5479063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5479209Z unimplemented [] 2025-12-04T09:45:06.5479319Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5479451Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5479554Z graph_break [] 2025-12-04T09:45:06.5479680Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5480021Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5480327Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5480503Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5480690Z Traceback (most recent call last): 2025-12-04T09:45:06.5480886Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5481078Z return value(self) 2025-12-04T09:45:06.5481288Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5481522Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5481744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5481976Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5482230Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5482529Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5482687Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5482775Z 2025-12-04T09:45:06.5482819Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5482997Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5483234Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5483371Z 2025-12-04T09:45:06.5483445Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5483740Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5483957Z 2025-12-04T09:45:06.5484046Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5484246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5484393Z unimplemented [] 2025-12-04T09:45:06.5484506Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5484640Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5484746Z graph_break [] 2025-12-04T09:45:06.5488857Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5489226Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5489546Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5489699Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5489849Z unimplemented [] 2025-12-04T09:45:06.5489964Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5490157Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5490271Z graph_break [] 2025-12-04T09:45:06.5490400Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5490741Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5491051Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5491195Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5491343Z unimplemented [] 2025-12-04T09:45:06.5491454Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5491588Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5491695Z graph_break [] 2025-12-04T09:45:06.5491820Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5492160Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5492469Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5492613Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5492759Z unimplemented [] 2025-12-04T09:45:06.5492868Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5493001Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5493106Z graph_break [] 2025-12-04T09:45:06.5493231Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5493567Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5493870Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5494014Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5494164Z unimplemented [] 2025-12-04T09:45:06.5494313Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5494447Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5494553Z graph_break [] 2025-12-04T09:45:06.5494678Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5495017Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5495324Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5495470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5495617Z unimplemented [] 2025-12-04T09:45:06.5495729Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5495863Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5495969Z graph_break [] 2025-12-04T09:45:06.5496095Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5496442Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5496794Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5496938Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5497086Z unimplemented [] 2025-12-04T09:45:06.5497195Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5497327Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5497434Z graph_break [] 2025-12-04T09:45:06.5497558Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5497894Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5498237Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5498381Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5498535Z unimplemented [] 2025-12-04T09:45:06.5498645Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5498776Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5498881Z graph_break [] 2025-12-04T09:45:06.5499006Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5499345Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5499649Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5499793Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5499939Z unimplemented [] 2025-12-04T09:45:06.5500050Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5500187Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5500292Z graph_break [] 2025-12-04T09:45:06.5500420Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5500759Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5501063Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5501205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5501353Z unimplemented [] 2025-12-04T09:45:06.5501464Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5501597Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5501702Z graph_break [] 2025-12-04T09:45:06.5501827Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5502165Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5502507Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5502650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5502796Z unimplemented [] 2025-12-04T09:45:06.5502906Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5503041Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5503147Z graph_break [] 2025-12-04T09:45:06.5503272Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5503612Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5503915Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5504058Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5504210Z unimplemented [] 2025-12-04T09:45:06.5504319Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5504578Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5504686Z graph_break [] 2025-12-04T09:45:06.5504813Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5505149Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5505454Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5505599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5505751Z unimplemented [] 2025-12-04T09:45:06.5505863Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5505997Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5506102Z graph_break [] 2025-12-04T09:45:06.5506263Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5506604Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5506950Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5507099Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5507247Z unimplemented [] 2025-12-04T09:45:06.5507360Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5507495Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5507604Z graph_break [] 2025-12-04T09:45:06.5507730Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5508072Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5508387Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5508533Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5508686Z unimplemented [] 2025-12-04T09:45:06.5508801Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5508937Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5509046Z graph_break [] 2025-12-04T09:45:06.5509171Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5509510Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5509815Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5509959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5510107Z unimplemented [] 2025-12-04T09:45:06.5510220Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5510358Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5510468Z graph_break [] 2025-12-04T09:45:06.5510634Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5510975Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5511283Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5511429Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5511577Z unimplemented [] 2025-12-04T09:45:06.5511691Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5511827Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5511937Z graph_break [] 2025-12-04T09:45:06.5512070Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5512418Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5512731Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5512878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5513031Z unimplemented [] 2025-12-04T09:45:06.5513143Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5513282Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5513393Z graph_break [] 2025-12-04T09:45:06.5513526Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5513871Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5514179Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5514325Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5514512Z unimplemented [] 2025-12-04T09:45:06.5514632Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5514777Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5514890Z graph_break [] 2025-12-04T09:45:06.5515027Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5515369Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5515680Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5515828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5515976Z unimplemented [] 2025-12-04T09:45:06.5516090Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5516227Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5516337Z graph_break [] 2025-12-04T09:45:06.5516471Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5516850Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5517161Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5517308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5517457Z unimplemented [] 2025-12-04T09:45:06.5517570Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5517709Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5517818Z graph_break [] 2025-12-04T09:45:06.5517949Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5518293Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5518600Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5518751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5518948Z unimplemented [] 2025-12-04T09:45:06.5519068Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5519205Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5519315Z graph_break [] 2025-12-04T09:45:06.5519447Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5519793Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5520103Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5520257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5520405Z unimplemented [] 2025-12-04T09:45:06.5520518Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5520654Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5520765Z graph_break [] 2025-12-04T09:45:06.5520899Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5521240Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5521552Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5521702Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5521851Z unimplemented [] 2025-12-04T09:45:06.5521965Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5522099Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5522205Z graph_break [] 2025-12-04T09:45:06.5522334Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5522703Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5523017Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5523197Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5523388Z Traceback (most recent call last): 2025-12-04T09:45:06.5523590Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5523785Z return value(self) 2025-12-04T09:45:06.5523999Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5524234Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5524462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5524698Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5524956Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5525227Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5525281Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5525286Z 2025-12-04T09:45:06.5525331Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5525439Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5525545Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5525548Z 2025-12-04T09:45:06.5525624Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5525808Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5525810Z 2025-12-04T09:45:06.5525902Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5525980Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5526050Z unimplemented [] 2025-12-04T09:45:06.5526108Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5526155Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5526192Z graph_break [] 2025-12-04T09:45:06.5526266Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5526501Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5526548Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5526620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5526661Z unimplemented [] 2025-12-04T09:45:06.5526719Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5526800Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5526837Z graph_break [] 2025-12-04T09:45:06.5526914Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5527146Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5527191Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5527265Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5527305Z unimplemented [] 2025-12-04T09:45:06.5527360Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5527403Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5527439Z graph_break [] 2025-12-04T09:45:06.5527511Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5527769Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5527814Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5527889Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5527929Z unimplemented [] 2025-12-04T09:45:06.5527986Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5528032Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5528067Z graph_break [] 2025-12-04T09:45:06.5528142Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5528369Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5528417Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5528490Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5528531Z unimplemented [] 2025-12-04T09:45:06.5528591Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5528640Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5528677Z graph_break [] 2025-12-04T09:45:06.5528752Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5528980Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5529026Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5529098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5529139Z unimplemented [] 2025-12-04T09:45:06.5529194Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5529239Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5529277Z graph_break [] 2025-12-04T09:45:06.5529353Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5529587Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5529662Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5529734Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5529773Z unimplemented [] 2025-12-04T09:45:06.5529828Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5529873Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5529908Z graph_break [] 2025-12-04T09:45:06.5529983Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5530209Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5530254Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5530328Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5530368Z unimplemented [] 2025-12-04T09:45:06.5530423Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5530467Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5530502Z graph_break [] 2025-12-04T09:45:06.5530576Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5530802Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5530844Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5530916Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5530954Z unimplemented [] 2025-12-04T09:45:06.5531008Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5531053Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5531089Z graph_break [] 2025-12-04T09:45:06.5531189Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5531418Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5531463Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5531537Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5531576Z unimplemented [] 2025-12-04T09:45:06.5531631Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5531675Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5531711Z graph_break [] 2025-12-04T09:45:06.5531785Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5532014Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5532057Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5532133Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5532170Z unimplemented [] 2025-12-04T09:45:06.5532226Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5532270Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5532305Z graph_break [] 2025-12-04T09:45:06.5532378Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5532606Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5532647Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5532721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5532758Z unimplemented [] 2025-12-04T09:45:06.5532815Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5532857Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5532923Z graph_break [] 2025-12-04T09:45:06.5532995Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5533223Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5533264Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5533337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5533374Z unimplemented [] 2025-12-04T09:45:06.5533430Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5533472Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5533511Z graph_break [] 2025-12-04T09:45:06.5533586Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5533818Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5533862Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5533935Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5533972Z unimplemented [] 2025-12-04T09:45:06.5534030Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5534072Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5534111Z graph_break [] 2025-12-04T09:45:06.5534183Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5534411Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5534452Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5534549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5534586Z unimplemented [] 2025-12-04T09:45:06.5534645Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5534687Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5534724Z graph_break [] 2025-12-04T09:45:06.5534797Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5535031Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5535073Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5535148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5535184Z unimplemented [] 2025-12-04T09:45:06.5535241Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5535283Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5535320Z graph_break [] 2025-12-04T09:45:06.5535393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5535623Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5535664Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5535738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5535775Z unimplemented [] 2025-12-04T09:45:06.5535831Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5535874Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5535911Z graph_break [] 2025-12-04T09:45:06.5535983Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5536211Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5536256Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5536367Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5536405Z unimplemented [] 2025-12-04T09:45:06.5536460Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5536503Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5536539Z graph_break [] 2025-12-04T09:45:06.5536611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5536873Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5536915Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5536991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5537028Z unimplemented [] 2025-12-04T09:45:06.5537084Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5537128Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5537167Z graph_break [] 2025-12-04T09:45:06.5537239Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5537466Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5537508Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5537581Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5537618Z unimplemented [] 2025-12-04T09:45:06.5537675Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5537717Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5537754Z graph_break [] 2025-12-04T09:45:06.5537829Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5538087Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5538133Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5538207Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5538244Z unimplemented [] 2025-12-04T09:45:06.5538300Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5538342Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5538378Z graph_break [] 2025-12-04T09:45:06.5538450Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5538677Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5538718Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5538797Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5538834Z unimplemented [] 2025-12-04T09:45:06.5538892Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5538934Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5538971Z graph_break [] 2025-12-04T09:45:06.5539044Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5539275Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5539317Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5539390Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5539427Z unimplemented [] 2025-12-04T09:45:06.5539483Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5539525Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5539563Z graph_break [] 2025-12-04T09:45:06.5539637Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5539902Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5539944Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5540018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5540055Z unimplemented [] 2025-12-04T09:45:06.5540110Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5540152Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5540190Z graph_break [] 2025-12-04T09:45:06.5540262Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5540491Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5540536Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5540610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5540648Z unimplemented [] 2025-12-04T09:45:06.5540703Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5540745Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5540782Z graph_break [] 2025-12-04T09:45:06.5540854Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5541085Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5541128Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5541233Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5541280Z Traceback (most recent call last): 2025-12-04T09:45:06.5541424Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5541467Z return value(self) 2025-12-04T09:45:06.5541615Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5541668Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5541806Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5541868Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5542027Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5542101Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5542156Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5542158Z 2025-12-04T09:45:06.5542203Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5542307Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5542413Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5542415Z 2025-12-04T09:45:06.5542488Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5542673Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5542675Z 2025-12-04T09:45:06.5542761Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5542836Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5542874Z unimplemented [] 2025-12-04T09:45:06.5542931Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5542974Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5543012Z graph_break [] 2025-12-04T09:45:06.5543086Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5543320Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5543393Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5543468Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5543505Z unimplemented [] 2025-12-04T09:45:06.5543562Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5543604Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5543642Z graph_break [] 2025-12-04T09:45:06.5543714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5543943Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5543986Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5544060Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5544099Z unimplemented [] 2025-12-04T09:45:06.5544156Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5544198Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5544235Z graph_break [] 2025-12-04T09:45:06.5544307Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5544536Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5544579Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5544651Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5544688Z unimplemented [] 2025-12-04T09:45:06.5544744Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5544821Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5544860Z graph_break [] 2025-12-04T09:45:06.5544934Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5545166Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5545209Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5545281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5545319Z unimplemented [] 2025-12-04T09:45:06.5545374Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5545416Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5545453Z graph_break [] 2025-12-04T09:45:06.5545525Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5545827Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5545873Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5545946Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5545984Z unimplemented [] 2025-12-04T09:45:06.5546039Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5546083Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5546118Z graph_break [] 2025-12-04T09:45:06.5546191Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5546416Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5546461Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5546534Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5546574Z unimplemented [] 2025-12-04T09:45:06.5546629Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5546706Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5546778Z graph_break [] 2025-12-04T09:45:06.5546852Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5547083Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5547126Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5547197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5547235Z unimplemented [] 2025-12-04T09:45:06.5547290Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5547334Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5547370Z graph_break [] 2025-12-04T09:45:06.5547445Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5547669Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5547714Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5547786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5547828Z unimplemented [] 2025-12-04T09:45:06.5547883Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5547927Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5547963Z graph_break [] 2025-12-04T09:45:06.5548036Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5548261Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5548305Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5548414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5548456Z unimplemented [] 2025-12-04T09:45:06.5548511Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5548556Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5548591Z graph_break [] 2025-12-04T09:45:06.5548665Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5548892Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5548936Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5549007Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5549046Z unimplemented [] 2025-12-04T09:45:06.5549101Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5549146Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5549184Z graph_break [] 2025-12-04T09:45:06.5549260Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5549486Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5549531Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5549603Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5549642Z unimplemented [] 2025-12-04T09:45:06.5549696Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5549740Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5549776Z graph_break [] 2025-12-04T09:45:06.5549849Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5550078Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5550157Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5550229Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5550268Z unimplemented [] 2025-12-04T09:45:06.5550322Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5550366Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5550402Z graph_break [] 2025-12-04T09:45:06.5550476Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5550705Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5550749Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5550821Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5550859Z unimplemented [] 2025-12-04T09:45:06.5550915Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5550960Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5550996Z graph_break [] 2025-12-04T09:45:06.5551070Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5551298Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5551342Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5551414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5551454Z unimplemented [] 2025-12-04T09:45:06.5551509Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5551553Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5551588Z graph_break [] 2025-12-04T09:45:06.5551663Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5551911Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5551958Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5552030Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5552068Z unimplemented [] 2025-12-04T09:45:06.5552123Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5552167Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5552204Z graph_break [] 2025-12-04T09:45:06.5552277Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5552503Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5552546Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5552620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5552663Z unimplemented [] 2025-12-04T09:45:06.5552718Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5552762Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5552797Z graph_break [] 2025-12-04T09:45:06.5552871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5553097Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5553141Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5553212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5553251Z unimplemented [] 2025-12-04T09:45:06.5553305Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5553350Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5553387Z graph_break [] 2025-12-04T09:45:06.5553461Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5553715Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5553759Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5553832Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5553869Z unimplemented [] 2025-12-04T09:45:06.5553924Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5553969Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5554003Z graph_break [] 2025-12-04T09:45:06.5554078Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5554308Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5554352Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5554425Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5554463Z unimplemented [] 2025-12-04T09:45:06.5554518Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5554561Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5554598Z graph_break [] 2025-12-04T09:45:06.5554671Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5554901Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5554943Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5555017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5555053Z unimplemented [] 2025-12-04T09:45:06.5555132Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5555177Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5555214Z graph_break [] 2025-12-04T09:45:06.5555287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5555515Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5555556Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5555629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5555666Z unimplemented [] 2025-12-04T09:45:06.5555724Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5555766Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5555803Z graph_break [] 2025-12-04T09:45:06.5555875Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5556105Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5556150Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5556226Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5556262Z unimplemented [] 2025-12-04T09:45:06.5556318Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5556360Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5556397Z graph_break [] 2025-12-04T09:45:06.5556469Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5556697Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5556772Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5556850Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5556925Z unimplemented [] 2025-12-04T09:45:06.5556982Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5557026Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5557063Z graph_break [] 2025-12-04T09:45:06.5557136Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5557363Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5557404Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5557478Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5557515Z unimplemented [] 2025-12-04T09:45:06.5557572Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5557614Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5557651Z graph_break [] 2025-12-04T09:45:06.5557725Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5557955Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5557997Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5558070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5558109Z unimplemented [] 2025-12-04T09:45:06.5558165Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5558206Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5558243Z graph_break [] 2025-12-04T09:45:06.5558316Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5558572Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5558614Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5558721Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5558766Z Traceback (most recent call last): 2025-12-04T09:45:06.5558886Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5558926Z return value(self) 2025-12-04T09:45:06.5559078Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5559129Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5559268Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5559327Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5559491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5559563Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5559619Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5559621Z 2025-12-04T09:45:06.5559663Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5559766Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5559868Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5559872Z 2025-12-04T09:45:06.5559945Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5560128Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5560130Z 2025-12-04T09:45:06.5560217Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5560294Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5560332Z unimplemented [] 2025-12-04T09:45:06.5560419Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5560462Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5560500Z graph_break [] 2025-12-04T09:45:06.5560573Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5560805Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5560847Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5560921Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5560959Z unimplemented [] 2025-12-04T09:45:06.5561020Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5561065Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5561104Z graph_break [] 2025-12-04T09:45:06.5561180Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5561418Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5561462Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5561541Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5561580Z unimplemented [] 2025-12-04T09:45:06.5561642Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5561686Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5561728Z graph_break [] 2025-12-04T09:45:06.5561800Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5562033Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5562109Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5562188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5562227Z unimplemented [] 2025-12-04T09:45:06.5562290Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5562332Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5562376Z graph_break [] 2025-12-04T09:45:06.5562451Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5562680Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5562724Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5562800Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5562838Z unimplemented [] 2025-12-04T09:45:06.5562899Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5562944Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5562987Z graph_break [] 2025-12-04T09:45:06.5563060Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5563292Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5563338Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5563417Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5563456Z unimplemented [] 2025-12-04T09:45:06.5563514Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5563558Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5563598Z graph_break [] 2025-12-04T09:45:06.5563671Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5563908Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5563980Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5564057Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5564100Z unimplemented [] 2025-12-04T09:45:06.5564172Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5564224Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5564270Z graph_break [] 2025-12-04T09:45:06.5564343Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5564573Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5564616Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5564695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5564735Z unimplemented [] 2025-12-04T09:45:06.5564797Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5564842Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5564885Z graph_break [] 2025-12-04T09:45:06.5564958Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5565186Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5565230Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5565306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5565345Z unimplemented [] 2025-12-04T09:45:06.5565404Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5565447Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5565490Z graph_break [] 2025-12-04T09:45:06.5565585Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5565818Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5565861Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5565936Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5565975Z unimplemented [] 2025-12-04T09:45:06.5566035Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5566079Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5566120Z graph_break [] 2025-12-04T09:45:06.5566193Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5566424Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5566470Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5566549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5566587Z unimplemented [] 2025-12-04T09:45:06.5566647Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5566691Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5566731Z graph_break [] 2025-12-04T09:45:06.5566890Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5567120Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5567165Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5567240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5567281Z unimplemented [] 2025-12-04T09:45:06.5567341Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5567386Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5567461Z graph_break [] 2025-12-04T09:45:06.5567536Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5567766Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5567812Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5567886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5567927Z unimplemented [] 2025-12-04T09:45:06.5567982Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5568028Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5568066Z graph_break [] 2025-12-04T09:45:06.5568140Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5568378Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5568427Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5568501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5568543Z unimplemented [] 2025-12-04T09:45:06.5568600Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5568648Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5568685Z graph_break [] 2025-12-04T09:45:06.5568762Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5568990Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5569038Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5569112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5569183Z unimplemented [] 2025-12-04T09:45:06.5569243Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5569291Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5569330Z graph_break [] 2025-12-04T09:45:06.5569407Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5569636Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5569682Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5569755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5569797Z unimplemented [] 2025-12-04T09:45:06.5569853Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5569900Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5569937Z graph_break [] 2025-12-04T09:45:06.5570017Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5570247Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5570293Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5570366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5570409Z unimplemented [] 2025-12-04T09:45:06.5570467Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5570513Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5570554Z graph_break [] 2025-12-04T09:45:06.5570632Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5570860Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5570908Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5570982Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5571054Z unimplemented [] 2025-12-04T09:45:06.5571112Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5571159Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5571196Z graph_break [] 2025-12-04T09:45:06.5571274Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5571500Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5571548Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5571621Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5571666Z unimplemented [] 2025-12-04T09:45:06.5571723Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5571773Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5571813Z graph_break [] 2025-12-04T09:45:06.5571891Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5572118Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5572164Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5572238Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5572279Z unimplemented [] 2025-12-04T09:45:06.5572334Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5572384Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5572421Z graph_break [] 2025-12-04T09:45:06.5572499Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5572752Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5572803Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5572877Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5572919Z unimplemented [] 2025-12-04T09:45:06.5572975Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5573023Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5573060Z graph_break [] 2025-12-04T09:45:06.5573138Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5573366Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5573413Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5573486Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5573531Z unimplemented [] 2025-12-04T09:45:06.5573587Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5573636Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5573674Z graph_break [] 2025-12-04T09:45:06.5573750Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5573977Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5574025Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5574098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5574143Z unimplemented [] 2025-12-04T09:45:06.5574200Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5574246Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5574282Z graph_break [] 2025-12-04T09:45:06.5574357Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5574587Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5574669Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5574741Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5574780Z unimplemented [] 2025-12-04T09:45:06.5574836Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5574879Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5574916Z graph_break [] 2025-12-04T09:45:06.5574990Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5575216Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5575261Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5575333Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5575376Z unimplemented [] 2025-12-04T09:45:06.5575432Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5575476Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5575512Z graph_break [] 2025-12-04T09:45:06.5575586Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5575812Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5575856Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5575929Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5575969Z unimplemented [] 2025-12-04T09:45:06.5576023Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5576088Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5576124Z graph_break [] 2025-12-04T09:45:06.5576201Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5576429Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5576473Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5576548Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5576585Z unimplemented [] 2025-12-04T09:45:06.5576642Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5576684Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5576722Z graph_break [] 2025-12-04T09:45:06.5576836Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5577065Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5577108Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5577213Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5577259Z Traceback (most recent call last): 2025-12-04T09:45:06.5577379Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5577418Z return value(self) 2025-12-04T09:45:06.5577567Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5577619Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5577757Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5577816Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5577977Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5578089Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5578144Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5578146Z 2025-12-04T09:45:06.5578192Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5578295Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5578397Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5578399Z 2025-12-04T09:45:06.5578474Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5578657Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5578658Z 2025-12-04T09:45:06.5578747Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5578823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5578865Z unimplemented [] 2025-12-04T09:45:06.5578923Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5578967Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5579003Z graph_break [] 2025-12-04T09:45:06.5579078Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5579307Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5579352Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5579424Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5579464Z unimplemented [] 2025-12-04T09:45:06.5579519Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5579564Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5579628Z graph_break [] 2025-12-04T09:45:06.5579704Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5579933Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5579977Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5580050Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5580089Z unimplemented [] 2025-12-04T09:45:06.5580144Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5580188Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5580224Z graph_break [] 2025-12-04T09:45:06.5580297Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5580527Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5580573Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5580647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5580686Z unimplemented [] 2025-12-04T09:45:06.5580741Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5580785Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5580820Z graph_break [] 2025-12-04T09:45:06.5580896Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5581125Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5581169Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5581243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5581280Z unimplemented [] 2025-12-04T09:45:06.5581338Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5581408Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5581443Z graph_break [] 2025-12-04T09:45:06.5581518Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5581745Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5581787Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5581863Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5581899Z unimplemented [] 2025-12-04T09:45:06.5581955Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5581997Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5582035Z graph_break [] 2025-12-04T09:45:06.5582107Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5582338Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5582381Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5582455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5582492Z unimplemented [] 2025-12-04T09:45:06.5582549Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5582591Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5582629Z graph_break [] 2025-12-04T09:45:06.5582701Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5582928Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5582973Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5583073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5583113Z unimplemented [] 2025-12-04T09:45:06.5583170Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5583213Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5583251Z graph_break [] 2025-12-04T09:45:06.5583323Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5583551Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5583593Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5583668Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5583706Z unimplemented [] 2025-12-04T09:45:06.5583762Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5583804Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5583841Z graph_break [] 2025-12-04T09:45:06.5583914Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5584144Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5584186Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5584260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5584297Z unimplemented [] 2025-12-04T09:45:06.5584353Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5584396Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5584433Z graph_break [] 2025-12-04T09:45:06.5584505Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5584735Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5584777Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5584878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5584915Z unimplemented [] 2025-12-04T09:45:06.5584971Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5585013Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5585050Z graph_break [] 2025-12-04T09:45:06.5585124Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5585352Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5585393Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5585467Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5585504Z unimplemented [] 2025-12-04T09:45:06.5585562Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5585605Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5585642Z graph_break [] 2025-12-04T09:45:06.5585714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5585941Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5585982Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5586056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5586093Z unimplemented [] 2025-12-04T09:45:06.5586149Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5586191Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5586228Z graph_break [] 2025-12-04T09:45:06.5586300Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5586551Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5586595Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5586671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5586708Z unimplemented [] 2025-12-04T09:45:06.5586793Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5586836Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5586874Z graph_break [] 2025-12-04T09:45:06.5586946Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5587178Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5587219Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5587296Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5587334Z unimplemented [] 2025-12-04T09:45:06.5587391Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5587436Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5587473Z graph_break [] 2025-12-04T09:45:06.5587546Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5587773Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5587816Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5587890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5587928Z unimplemented [] 2025-12-04T09:45:06.5587985Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5588027Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5588064Z graph_break [] 2025-12-04T09:45:06.5588139Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5588410Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5588451Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5588525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5588563Z unimplemented [] 2025-12-04T09:45:06.5588620Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5588663Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5588701Z graph_break [] 2025-12-04T09:45:06.5588773Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5589006Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5589048Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5589125Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5589163Z unimplemented [] 2025-12-04T09:45:06.5589219Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5589262Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5589299Z graph_break [] 2025-12-04T09:45:06.5589371Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5589600Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5589644Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5589716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5589755Z unimplemented [] 2025-12-04T09:45:06.5589837Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5589881Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5589920Z graph_break [] 2025-12-04T09:45:06.5589993Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5590221Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5590264Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5590336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5590375Z unimplemented [] 2025-12-04T09:45:06.5590430Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5590475Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5590510Z graph_break [] 2025-12-04T09:45:06.5590584Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5590813Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5590859Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5590931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5590970Z unimplemented [] 2025-12-04T09:45:06.5591025Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5591070Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5591105Z graph_break [] 2025-12-04T09:45:06.5591179Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5591407Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5591452Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5591526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5591565Z unimplemented [] 2025-12-04T09:45:06.5591650Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5591693Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5591729Z graph_break [] 2025-12-04T09:45:06.5591803Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5592030Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5592075Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5592149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5592188Z unimplemented [] 2025-12-04T09:45:06.5592243Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5592287Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5592323Z graph_break [] 2025-12-04T09:45:06.5592399Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5592631Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5592674Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5592747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5592787Z unimplemented [] 2025-12-04T09:45:06.5592842Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5592886Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5592922Z graph_break [] 2025-12-04T09:45:06.5593000Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5593230Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5593297Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5593373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5593413Z unimplemented [] 2025-12-04T09:45:06.5593467Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5593512Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5593547Z graph_break [] 2025-12-04T09:45:06.5593624Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5593854Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5593898Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5593971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5594010Z unimplemented [] 2025-12-04T09:45:06.5594066Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5594111Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5594148Z graph_break [] 2025-12-04T09:45:06.5594221Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5594448Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5594491Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5594566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5594605Z unimplemented [] 2025-12-04T09:45:06.5594659Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5594704Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5594739Z graph_break [] 2025-12-04T09:45:06.5594814Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5595042Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5595113Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5595186Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5595225Z unimplemented [] 2025-12-04T09:45:06.5595281Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5595324Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5595361Z graph_break [] 2025-12-04T09:45:06.5595436Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5595662Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5595705Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5595811Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5595859Z Traceback (most recent call last): 2025-12-04T09:45:06.5595978Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5596019Z return value(self) 2025-12-04T09:45:06.5596167Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5596221Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5596357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5596419Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5596577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5596651Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5596704Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5596732Z 2025-12-04T09:45:06.5596820Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5596924Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5597030Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5597032Z 2025-12-04T09:45:06.5597107Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5597292Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5597294Z 2025-12-04T09:45:06.5597384Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5597458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5597501Z unimplemented [] 2025-12-04T09:45:06.5597560Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5597607Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5597645Z graph_break [] 2025-12-04T09:45:06.5597724Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5597956Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5598004Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5598078Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5598121Z unimplemented [] 2025-12-04T09:45:06.5598176Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5598222Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5598260Z graph_break [] 2025-12-04T09:45:06.5598335Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5598564Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5598647Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5598719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5598757Z unimplemented [] 2025-12-04T09:45:06.5598814Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5598858Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5598894Z graph_break [] 2025-12-04T09:45:06.5598968Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5599194Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5599239Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5599312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5599354Z unimplemented [] 2025-12-04T09:45:06.5599408Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5599456Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5599493Z graph_break [] 2025-12-04T09:45:06.5599568Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5599795Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5599838Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5599911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5599950Z unimplemented [] 2025-12-04T09:45:06.5600005Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5600049Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5600085Z graph_break [] 2025-12-04T09:45:06.5600206Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5600433Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5600480Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5600552Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5600591Z unimplemented [] 2025-12-04T09:45:06.5600647Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5600699Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5603148Z graph_break [] 2025-12-04T09:45:06.5603231Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5603463Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5603513Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5603586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5603629Z unimplemented [] 2025-12-04T09:45:06.5603684Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5603728Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5603764Z graph_break [] 2025-12-04T09:45:06.5603839Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5604068Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5604113Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5604185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5604224Z unimplemented [] 2025-12-04T09:45:06.5604279Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5604327Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5604363Z graph_break [] 2025-12-04T09:45:06.5604477Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5604707Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5604755Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5604827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5604866Z unimplemented [] 2025-12-04T09:45:06.5604922Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5604968Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5605006Z graph_break [] 2025-12-04T09:45:06.5605080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5605308Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5605356Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5605428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5605469Z unimplemented [] 2025-12-04T09:45:06.5605526Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5605572Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5605608Z graph_break [] 2025-12-04T09:45:06.5605681Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5605907Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5606356Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5606429Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5606493Z unimplemented [] 2025-12-04T09:45:06.5606551Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5606599Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5606636Z graph_break [] 2025-12-04T09:45:06.5606708Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5606972Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5607014Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5607091Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5607127Z unimplemented [] 2025-12-04T09:45:06.5607183Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5607225Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5607262Z graph_break [] 2025-12-04T09:45:06.5607338Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5607567Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5607610Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5607685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5607722Z unimplemented [] 2025-12-04T09:45:06.5607778Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5607821Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5607858Z graph_break [] 2025-12-04T09:45:06.5607932Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5608159Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5608201Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5608282Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5608359Z unimplemented [] 2025-12-04T09:45:06.5608415Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5608457Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5608494Z graph_break [] 2025-12-04T09:45:06.5608567Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5608794Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5608835Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5608909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5608945Z unimplemented [] 2025-12-04T09:45:06.5609001Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5609043Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5609085Z graph_break [] 2025-12-04T09:45:06.5609158Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5609385Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5609426Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5609501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5609537Z unimplemented [] 2025-12-04T09:45:06.5609594Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5609637Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5609676Z graph_break [] 2025-12-04T09:45:06.5609748Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5610009Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5610054Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5610129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5610165Z unimplemented [] 2025-12-04T09:45:06.5610222Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5610264Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5610300Z graph_break [] 2025-12-04T09:45:06.5610372Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5610600Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5610642Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5610715Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5610752Z unimplemented [] 2025-12-04T09:45:06.5610810Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5610855Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5610893Z graph_break [] 2025-12-04T09:45:06.5610966Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5611195Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5611236Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5611312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5611350Z unimplemented [] 2025-12-04T09:45:06.5611405Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5611448Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5611485Z graph_break [] 2025-12-04T09:45:06.5611556Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5611786Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5611860Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5611933Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5611970Z unimplemented [] 2025-12-04T09:45:06.5612025Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5612067Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5612105Z graph_break [] 2025-12-04T09:45:06.5612176Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5612405Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5612446Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5612522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5612561Z unimplemented [] 2025-12-04T09:45:06.5612617Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5612658Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5612696Z graph_break [] 2025-12-04T09:45:06.5612768Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5612998Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5613040Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5613116Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5613153Z unimplemented [] 2025-12-04T09:45:06.5613210Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5613251Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5613313Z graph_break [] 2025-12-04T09:45:06.5613386Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5613618Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5613661Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5613735Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5613772Z unimplemented [] 2025-12-04T09:45:06.5613829Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5613871Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5613908Z graph_break [] 2025-12-04T09:45:06.5613980Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5614208Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5614252Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5614325Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5614363Z unimplemented [] 2025-12-04T09:45:06.5614421Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5614462Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5614501Z graph_break [] 2025-12-04T09:45:06.5614573Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5614803Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5614845Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5614918Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5614956Z unimplemented [] 2025-12-04T09:45:06.5615012Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5615087Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5615122Z graph_break [] 2025-12-04T09:45:06.5615194Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5615427Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5615470Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5615541Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5615582Z unimplemented [] 2025-12-04T09:45:06.5615637Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5615680Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5615717Z graph_break [] 2025-12-04T09:45:06.5615793Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5616020Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5616064Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5616136Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5616175Z unimplemented [] 2025-12-04T09:45:06.5616230Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5616275Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5616310Z graph_break [] 2025-12-04T09:45:06.5616386Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5616613Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5616658Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5616808Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5616850Z unimplemented [] 2025-12-04T09:45:06.5616905Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5616950Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5616987Z graph_break [] 2025-12-04T09:45:06.5617062Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5617292Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5617337Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5617411Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5617454Z unimplemented [] 2025-12-04T09:45:06.5617509Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5617554Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5617594Z graph_break [] 2025-12-04T09:45:06.5617669Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5617898Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5617942Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5618046Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5618097Z Traceback (most recent call last): 2025-12-04T09:45:06.5618221Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5618264Z return value(self) 2025-12-04T09:45:06.5618413Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5618468Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5618611Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5618710Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5618869Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5618945Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5618999Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5619002Z 2025-12-04T09:45:06.5619048Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5619151Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5619257Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5619259Z 2025-12-04T09:45:06.5619333Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5619524Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5619527Z 2025-12-04T09:45:06.5619617Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5619693Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5619733Z unimplemented [] 2025-12-04T09:45:06.5619790Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5619836Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5619872Z graph_break [] 2025-12-04T09:45:06.5619945Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5620177Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5620224Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5620321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5620362Z unimplemented [] 2025-12-04T09:45:06.5620419Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5620465Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5620502Z graph_break [] 2025-12-04T09:45:06.5620577Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5620804Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5620849Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5620923Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5620964Z unimplemented [] 2025-12-04T09:45:06.5621019Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5621062Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5621099Z graph_break [] 2025-12-04T09:45:06.5621175Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5621405Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5621450Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5621522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5621562Z unimplemented [] 2025-12-04T09:45:06.5621618Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5621663Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5621698Z graph_break [] 2025-12-04T09:45:06.5621778Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5622005Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5622052Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5622154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5622194Z unimplemented [] 2025-12-04T09:45:06.5622250Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5622298Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5622336Z graph_break [] 2025-12-04T09:45:06.5622416Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5622643Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5622690Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5622763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5622807Z unimplemented [] 2025-12-04T09:45:06.5622864Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5622912Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5622956Z graph_break [] 2025-12-04T09:45:06.5623035Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5623262Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5623311Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5623385Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5623426Z unimplemented [] 2025-12-04T09:45:06.5623483Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5623532Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5623571Z graph_break [] 2025-12-04T09:45:06.5623649Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5623899Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5623949Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5624022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5624063Z unimplemented [] 2025-12-04T09:45:06.5624120Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5624170Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5624208Z graph_break [] 2025-12-04T09:45:06.5624289Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5624518Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5624567Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5624642Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5624687Z unimplemented [] 2025-12-04T09:45:06.5624747Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5624796Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5624834Z graph_break [] 2025-12-04T09:45:06.5624910Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5625137Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5625185Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5625258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5625298Z unimplemented [] 2025-12-04T09:45:06.5625357Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5625401Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5625442Z graph_break [] 2025-12-04T09:45:06.5625518Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5625778Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5625821Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5625898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5625937Z unimplemented [] 2025-12-04T09:45:06.5625997Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5626041Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5626083Z graph_break [] 2025-12-04T09:45:06.5626159Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5626392Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5626436Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5626515Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5626554Z unimplemented [] 2025-12-04T09:45:06.5626612Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5626657Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5626698Z graph_break [] 2025-12-04T09:45:06.5626806Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5627036Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5627078Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5627157Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5627196Z unimplemented [] 2025-12-04T09:45:06.5627256Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5627332Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5627377Z graph_break [] 2025-12-04T09:45:06.5627452Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5627685Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5627727Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5627804Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5627843Z unimplemented [] 2025-12-04T09:45:06.5627902Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5627947Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5627986Z graph_break [] 2025-12-04T09:45:06.5628060Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5628301Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5628348Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5628426Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5628465Z unimplemented [] 2025-12-04T09:45:06.5628525Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5628571Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5628612Z graph_break [] 2025-12-04T09:45:06.5628688Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5628921Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5628963Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5629041Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5629082Z unimplemented [] 2025-12-04T09:45:06.5629185Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5629227Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5629264Z graph_break [] 2025-12-04T09:45:06.5629336Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5629564Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5629607Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5629680Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5629717Z unimplemented [] 2025-12-04T09:45:06.5629773Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5629814Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5629850Z graph_break [] 2025-12-04T09:45:06.5629924Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5630153Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5630194Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5630267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5630304Z unimplemented [] 2025-12-04T09:45:06.5630359Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5630402Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5630439Z graph_break [] 2025-12-04T09:45:06.5630510Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5630736Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5630798Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5630872Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5630911Z unimplemented [] 2025-12-04T09:45:06.5630966Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5631008Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5631045Z graph_break [] 2025-12-04T09:45:06.5631117Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5631345Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5631387Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5631460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5631496Z unimplemented [] 2025-12-04T09:45:06.5631550Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5631595Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5631632Z graph_break [] 2025-12-04T09:45:06.5631706Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5631934Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5631974Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5632047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5632083Z unimplemented [] 2025-12-04T09:45:06.5632138Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5632180Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5632216Z graph_break [] 2025-12-04T09:45:06.5632288Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5632518Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5632586Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5632660Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5632696Z unimplemented [] 2025-12-04T09:45:06.5632752Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5632793Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5632828Z graph_break [] 2025-12-04T09:45:06.5632900Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5633127Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5633168Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5633241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5633279Z unimplemented [] 2025-12-04T09:45:06.5633335Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5633378Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5633414Z graph_break [] 2025-12-04T09:45:06.5633486Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5633712Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5633753Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5633826Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5633862Z unimplemented [] 2025-12-04T09:45:06.5633918Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5633959Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5633996Z graph_break [] 2025-12-04T09:45:06.5634094Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5634322Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5634365Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5634438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5634475Z unimplemented [] 2025-12-04T09:45:06.5634530Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5634571Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5634607Z graph_break [] 2025-12-04T09:45:06.5634678Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5634904Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5634946Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5635018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5635057Z unimplemented [] 2025-12-04T09:45:06.5635112Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5635154Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5635192Z graph_break [] 2025-12-04T09:45:06.5635263Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5635489Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5635530Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5635603Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5635640Z unimplemented [] 2025-12-04T09:45:06.5635695Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5635739Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5635776Z graph_break [] 2025-12-04T09:45:06.5635876Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5636104Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5636146Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5636218Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5636255Z unimplemented [] 2025-12-04T09:45:06.5636310Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5636352Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5636388Z graph_break [] 2025-12-04T09:45:06.5636460Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5636697Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5636778Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5636853Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5636889Z unimplemented [] 2025-12-04T09:45:06.5636943Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5636985Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5637021Z graph_break [] 2025-12-04T09:45:06.5637093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5637320Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5637362Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5637435Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5637508Z unimplemented [] 2025-12-04T09:45:06.5637563Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5637608Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5637643Z graph_break [] 2025-12-04T09:45:06.5637714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5637942Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5637984Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5638088Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5638134Z Traceback (most recent call last): 2025-12-04T09:45:06.5638253Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5638292Z return value(self) 2025-12-04T09:45:06.5638441Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5638494Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5638630Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5638691Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5638852Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5638925Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5638977Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5638979Z 2025-12-04T09:45:06.5639024Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5639125Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5639227Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5639230Z 2025-12-04T09:45:06.5639302Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5639519Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5639522Z 2025-12-04T09:45:06.5639607Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5639682Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5639720Z unimplemented [] 2025-12-04T09:45:06.5639776Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5639818Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5639854Z graph_break [] 2025-12-04T09:45:06.5639926Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5640158Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5640202Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5640275Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5640312Z unimplemented [] 2025-12-04T09:45:06.5640368Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5640410Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5640447Z graph_break [] 2025-12-04T09:45:06.5640519Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5640751Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5640792Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5640867Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5640904Z unimplemented [] 2025-12-04T09:45:06.5640982Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5641026Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5641062Z graph_break [] 2025-12-04T09:45:06.5641134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5641365Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5641406Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5641479Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5641516Z unimplemented [] 2025-12-04T09:45:06.5641571Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5641613Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5641650Z graph_break [] 2025-12-04T09:45:06.5641721Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5641949Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5641992Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5642064Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5642103Z unimplemented [] 2025-12-04T09:45:06.5642157Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5642200Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5642235Z graph_break [] 2025-12-04T09:45:06.5642306Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5642533Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5642576Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5642650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5642719Z unimplemented [] 2025-12-04T09:45:06.5642774Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5642817Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5642852Z graph_break [] 2025-12-04T09:45:06.5642928Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5643157Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5643199Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5643270Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5643309Z unimplemented [] 2025-12-04T09:45:06.5643363Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5643406Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5643444Z graph_break [] 2025-12-04T09:45:06.5643517Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5643743Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5643785Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5643858Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5643896Z unimplemented [] 2025-12-04T09:45:06.5643950Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5643993Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5644028Z graph_break [] 2025-12-04T09:45:06.5644100Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5644350Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5644395Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5644467Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5644504Z unimplemented [] 2025-12-04T09:45:06.5644558Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5644601Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5644636Z graph_break [] 2025-12-04T09:45:06.5644708Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5644933Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5644975Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5645046Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5645083Z unimplemented [] 2025-12-04T09:45:06.5645138Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5645182Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5645217Z graph_break [] 2025-12-04T09:45:06.5645290Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5645514Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5645556Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5645627Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5645665Z unimplemented [] 2025-12-04T09:45:06.5645720Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5645762Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5645797Z graph_break [] 2025-12-04T09:45:06.5645869Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5646095Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5646164Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5646236Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5646273Z unimplemented [] 2025-12-04T09:45:06.5646326Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5646370Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5646405Z graph_break [] 2025-12-04T09:45:06.5646477Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5646702Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5646773Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5646846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5646885Z unimplemented [] 2025-12-04T09:45:06.5646940Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5646983Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5647018Z graph_break [] 2025-12-04T09:45:06.5647094Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5647321Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5647363Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5647435Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5647473Z unimplemented [] 2025-12-04T09:45:06.5647527Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5647569Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5647604Z graph_break [] 2025-12-04T09:45:06.5647707Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5647934Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5647976Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5648048Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5648085Z unimplemented [] 2025-12-04T09:45:06.5648139Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5648182Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5648217Z graph_break [] 2025-12-04T09:45:06.5648292Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5648520Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5648562Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5648635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5648672Z unimplemented [] 2025-12-04T09:45:06.5648727Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5648769Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5648805Z graph_break [] 2025-12-04T09:45:06.5648878Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5649108Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5649151Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5649223Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5649262Z unimplemented [] 2025-12-04T09:45:06.5649317Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5649360Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5649428Z graph_break [] 2025-12-04T09:45:06.5649501Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5649729Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5649771Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5649842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5649880Z unimplemented [] 2025-12-04T09:45:06.5649934Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5649977Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5650013Z graph_break [] 2025-12-04T09:45:06.5650086Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5650315Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5650359Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5650432Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5650468Z unimplemented [] 2025-12-04T09:45:06.5650522Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5650565Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5650600Z graph_break [] 2025-12-04T09:45:06.5650672Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5650899Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5650940Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5651037Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5651076Z unimplemented [] 2025-12-04T09:45:06.5651131Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5651173Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5651209Z graph_break [] 2025-12-04T09:45:06.5651281Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5651510Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5651550Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5651622Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5651658Z unimplemented [] 2025-12-04T09:45:06.5651713Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5651755Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5651791Z graph_break [] 2025-12-04T09:45:06.5651863Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5652090Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5652131Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5652202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5652239Z unimplemented [] 2025-12-04T09:45:06.5652293Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5652335Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5652370Z graph_break [] 2025-12-04T09:45:06.5652441Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5652667Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5652709Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5652812Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5652848Z unimplemented [] 2025-12-04T09:45:06.5652904Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5652945Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5652981Z graph_break [] 2025-12-04T09:45:06.5653053Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5653286Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5653327Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5653400Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5653437Z unimplemented [] 2025-12-04T09:45:06.5653493Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5653535Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5653572Z graph_break [] 2025-12-04T09:45:06.5653643Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5653871Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5653912Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5653985Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5654021Z unimplemented [] 2025-12-04T09:45:06.5654077Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5654118Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5654154Z graph_break [] 2025-12-04T09:45:06.5654225Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5654486Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5654529Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5654602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5654638Z unimplemented [] 2025-12-04T09:45:06.5654693Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5654735Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5654772Z graph_break [] 2025-12-04T09:45:06.5654844Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5655073Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5655115Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5655190Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5655226Z unimplemented [] 2025-12-04T09:45:06.5655284Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5655325Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5655362Z graph_break [] 2025-12-04T09:45:06.5655435Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5655666Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5655707Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5655783Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5655820Z unimplemented [] 2025-12-04T09:45:06.5655876Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5655919Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5655956Z graph_break [] 2025-12-04T09:45:06.5656030Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5664350Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5664391Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5664465Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5664503Z unimplemented [] 2025-12-04T09:45:06.5664559Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5664602Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5664639Z graph_break [] 2025-12-04T09:45:06.5664711Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5664938Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5664982Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5665057Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5665094Z unimplemented [] 2025-12-04T09:45:06.5665149Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5665191Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5665228Z graph_break [] 2025-12-04T09:45:06.5665299Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5665530Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5665571Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5665645Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5665682Z unimplemented [] 2025-12-04T09:45:06.5665737Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5665847Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5665884Z graph_break [] 2025-12-04T09:45:06.5665956Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5666184Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5666226Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5666331Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5666377Z Traceback (most recent call last): 2025-12-04T09:45:06.5666496Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5666536Z return value(self) 2025-12-04T09:45:06.5666684Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5666774Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5666911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5666974Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5667132Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5667205Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5667258Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5667260Z 2025-12-04T09:45:06.5667305Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5667406Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5667509Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5667511Z 2025-12-04T09:45:06.5667584Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5667771Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5667808Z 2025-12-04T09:45:06.5667896Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5667970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5668008Z unimplemented [] 2025-12-04T09:45:06.5668065Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5668108Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5668145Z graph_break [] 2025-12-04T09:45:06.5668217Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5668449Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5668493Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5668566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5668605Z unimplemented [] 2025-12-04T09:45:06.5668661Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5668704Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5668741Z graph_break [] 2025-12-04T09:45:06.5668813Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5669040Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5669082Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5669154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5669191Z unimplemented [] 2025-12-04T09:45:06.5669247Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5669314Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5669352Z graph_break [] 2025-12-04T09:45:06.5669427Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5669655Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5669698Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5669772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5669809Z unimplemented [] 2025-12-04T09:45:06.5669865Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5669907Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5669944Z graph_break [] 2025-12-04T09:45:06.5670016Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5670245Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5670290Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5670363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5670400Z unimplemented [] 2025-12-04T09:45:06.5670456Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5670498Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5670534Z graph_break [] 2025-12-04T09:45:06.5670606Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5670834Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5670876Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5670949Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5670988Z unimplemented [] 2025-12-04T09:45:06.5671044Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5671110Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5671148Z graph_break [] 2025-12-04T09:45:06.5671219Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5671447Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5671488Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5671561Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5671598Z unimplemented [] 2025-12-04T09:45:06.5671653Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5671695Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5671732Z graph_break [] 2025-12-04T09:45:06.5671805Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5672034Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5672078Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5672149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5672188Z unimplemented [] 2025-12-04T09:45:06.5672241Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5672283Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5672319Z graph_break [] 2025-12-04T09:45:06.5672390Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5672616Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5672658Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5672749Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5672789Z unimplemented [] 2025-12-04T09:45:06.5672844Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5672887Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5672922Z graph_break [] 2025-12-04T09:45:06.5672995Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5673222Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5673265Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5673336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5673374Z unimplemented [] 2025-12-04T09:45:06.5673428Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5673471Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5673508Z graph_break [] 2025-12-04T09:45:06.5673584Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5673810Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5673853Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5673926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5673964Z unimplemented [] 2025-12-04T09:45:06.5674018Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5674061Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5674097Z graph_break [] 2025-12-04T09:45:06.5674169Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5674398Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5674467Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5674539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5674576Z unimplemented [] 2025-12-04T09:45:06.5674631Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5674675Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5674710Z graph_break [] 2025-12-04T09:45:06.5674783Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5675009Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5675053Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5675124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5675162Z unimplemented [] 2025-12-04T09:45:06.5675217Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5675262Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5675298Z graph_break [] 2025-12-04T09:45:06.5675371Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5675595Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5675638Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5675710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5675749Z unimplemented [] 2025-12-04T09:45:06.5675803Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5675845Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5675880Z graph_break [] 2025-12-04T09:45:06.5675958Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5676217Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5676265Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5676337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5676377Z unimplemented [] 2025-12-04T09:45:06.5676432Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5676477Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5676513Z graph_break [] 2025-12-04T09:45:06.5676588Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5676853Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5676899Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5676977Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5677021Z unimplemented [] 2025-12-04T09:45:06.5677077Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5677122Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5677158Z graph_break [] 2025-12-04T09:45:06.5677234Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5677459Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5677505Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5677580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5677621Z unimplemented [] 2025-12-04T09:45:06.5677677Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5677722Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5677759Z graph_break [] 2025-12-04T09:45:06.5677835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5678097Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5678140Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5678213Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5678254Z unimplemented [] 2025-12-04T09:45:06.5678309Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5678354Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5678390Z graph_break [] 2025-12-04T09:45:06.5678465Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5678693Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5678740Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5678812Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5678854Z unimplemented [] 2025-12-04T09:45:06.5678909Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5678955Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5678990Z graph_break [] 2025-12-04T09:45:06.5679065Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5679292Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5679336Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5679409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5679449Z unimplemented [] 2025-12-04T09:45:06.5679535Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5679584Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5679620Z graph_break [] 2025-12-04T09:45:06.5679696Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5679921Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5679967Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5680039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5680079Z unimplemented [] 2025-12-04T09:45:06.5680134Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5680181Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5680217Z graph_break [] 2025-12-04T09:45:06.5680296Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5680526Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5680572Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5680647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5680685Z unimplemented [] 2025-12-04T09:45:06.5680742Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5680784Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5680820Z graph_break [] 2025-12-04T09:45:06.5680895Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5681124Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5681167Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5681243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5681307Z unimplemented [] 2025-12-04T09:45:06.5681367Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5681410Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5681449Z graph_break [] 2025-12-04T09:45:06.5681520Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5681749Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5681792Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5681867Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5681905Z unimplemented [] 2025-12-04T09:45:06.5681962Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5682005Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5682045Z graph_break [] 2025-12-04T09:45:06.5682119Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5682352Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5682395Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5682473Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5682511Z unimplemented [] 2025-12-04T09:45:06.5682568Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5682611Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5682652Z graph_break [] 2025-12-04T09:45:06.5682725Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5682975Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5683020Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5683097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5683136Z unimplemented [] 2025-12-04T09:45:06.5683194Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5683237Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5683275Z graph_break [] 2025-12-04T09:45:06.5683348Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5683576Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5683618Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5683693Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5683729Z unimplemented [] 2025-12-04T09:45:06.5683789Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5683833Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5683871Z graph_break [] 2025-12-04T09:45:06.5683944Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5684171Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5684212Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5684287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5684325Z unimplemented [] 2025-12-04T09:45:06.5684383Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5684425Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5684464Z graph_break [] 2025-12-04T09:45:06.5684539Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5684770Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5684843Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5684919Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5684957Z unimplemented [] 2025-12-04T09:45:06.5685014Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5685057Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5685097Z graph_break [] 2025-12-04T09:45:06.5685170Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5685398Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5685441Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5685517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5685557Z unimplemented [] 2025-12-04T09:45:06.5685614Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5685656Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5685696Z graph_break [] 2025-12-04T09:45:06.5685768Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5685998Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5686040Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5686114Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5686151Z unimplemented [] 2025-12-04T09:45:06.5686211Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5686254Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5686293Z graph_break [] 2025-12-04T09:45:06.5686388Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5686624Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5686666Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5686773Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5686811Z unimplemented [] 2025-12-04T09:45:06.5686870Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5686913Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5686952Z graph_break [] 2025-12-04T09:45:06.5687025Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5687256Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5687298Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5687408Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5687454Z Traceback (most recent call last): 2025-12-04T09:45:06.5687577Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5687617Z return value(self) 2025-12-04T09:45:06.5687768Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5687822Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5687962Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5688022Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5688185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5688259Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5688362Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5688365Z 2025-12-04T09:45:06.5688410Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5688513Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5688617Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5688620Z 2025-12-04T09:45:06.5688693Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5688880Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5688882Z 2025-12-04T09:45:06.5688970Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5689047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5689086Z unimplemented [] 2025-12-04T09:45:06.5689146Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5689190Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5689230Z graph_break [] 2025-12-04T09:45:06.5689304Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5689537Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5689580Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5689656Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5689694Z unimplemented [] 2025-12-04T09:45:06.5689753Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5689796Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5689836Z graph_break [] 2025-12-04T09:45:06.5689933Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5690169Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5690212Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5690287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5690327Z unimplemented [] 2025-12-04T09:45:06.5690386Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5690429Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5690471Z graph_break [] 2025-12-04T09:45:06.5690543Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5690773Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5690818Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5690896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5690933Z unimplemented [] 2025-12-04T09:45:06.5690993Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5691035Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5691077Z graph_break [] 2025-12-04T09:45:06.5691151Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5691381Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5691423Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5691498Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5691536Z unimplemented [] 2025-12-04T09:45:06.5691598Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5691642Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5691708Z graph_break [] 2025-12-04T09:45:06.5691781Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5692012Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5692054Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5692129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5692168Z unimplemented [] 2025-12-04T09:45:06.5692228Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5692271Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5692311Z graph_break [] 2025-12-04T09:45:06.5692383Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5692614Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5692660Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5692739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5692777Z unimplemented [] 2025-12-04T09:45:06.5692836Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5692880Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5692919Z graph_break [] 2025-12-04T09:45:06.5692992Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5693226Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5693269Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5693343Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5693403Z unimplemented [] 2025-12-04T09:45:06.5693464Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5693507Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5693547Z graph_break [] 2025-12-04T09:45:06.5693620Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5693848Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5693894Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5693969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5694009Z unimplemented [] 2025-12-04T09:45:06.5694065Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5694108Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5694145Z graph_break [] 2025-12-04T09:45:06.5694220Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5694450Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5694496Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5694569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5694612Z unimplemented [] 2025-12-04T09:45:06.5694667Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5694712Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5694748Z graph_break [] 2025-12-04T09:45:06.5694822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5695051Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5695097Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5695171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5695234Z unimplemented [] 2025-12-04T09:45:06.5695288Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5695331Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5695367Z graph_break [] 2025-12-04T09:45:06.5695440Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5695668Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5695714Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5695787Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5695829Z unimplemented [] 2025-12-04T09:45:06.5695883Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5695931Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5695967Z graph_break [] 2025-12-04T09:45:06.5696046Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5696273Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5696318Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5696391Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5696431Z unimplemented [] 2025-12-04T09:45:06.5696487Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5696532Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5696568Z graph_break [] 2025-12-04T09:45:06.5696645Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5696936Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5696984Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5697057Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5697097Z unimplemented [] 2025-12-04T09:45:06.5697153Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5697199Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5697235Z graph_break [] 2025-12-04T09:45:06.5697310Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5697540Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5697584Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5697656Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5697697Z unimplemented [] 2025-12-04T09:45:06.5697751Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5697797Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5697832Z graph_break [] 2025-12-04T09:45:06.5697905Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5698131Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5698174Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5698246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5698283Z unimplemented [] 2025-12-04T09:45:06.5698337Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5698380Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5698416Z graph_break [] 2025-12-04T09:45:06.5698491Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5698717Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5698787Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5698859Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5698897Z unimplemented [] 2025-12-04T09:45:06.5698951Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5698996Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5699031Z graph_break [] 2025-12-04T09:45:06.5699106Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5699332Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5699377Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5699450Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5699491Z unimplemented [] 2025-12-04T09:45:06.5699546Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5699590Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5699625Z graph_break [] 2025-12-04T09:45:06.5699698Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5699923Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5699966Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5700038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5700076Z unimplemented [] 2025-12-04T09:45:06.5700130Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5700195Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5700232Z graph_break [] 2025-12-04T09:45:06.5700307Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5700534Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5700578Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5700650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5700689Z unimplemented [] 2025-12-04T09:45:06.5700744Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5700787Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5700823Z graph_break [] 2025-12-04T09:45:06.5700897Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5701127Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5701172Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5701244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5701285Z unimplemented [] 2025-12-04T09:45:06.5701339Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5701383Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5701419Z graph_break [] 2025-12-04T09:45:06.5701498Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5701728Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5701770Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5701843Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5701886Z unimplemented [] 2025-12-04T09:45:06.5701941Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5702015Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5702051Z graph_break [] 2025-12-04T09:45:06.5702124Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5702350Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5702393Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5702467Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5702503Z unimplemented [] 2025-12-04T09:45:06.5702558Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5702602Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5702636Z graph_break [] 2025-12-04T09:45:06.5702714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5702942Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5702987Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5703060Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5703098Z unimplemented [] 2025-12-04T09:45:06.5703153Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5703195Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5703234Z graph_break [] 2025-12-04T09:45:06.5703306Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5703534Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5703575Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5703672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5703711Z unimplemented [] 2025-12-04T09:45:06.5703767Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5703809Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5703848Z graph_break [] 2025-12-04T09:45:06.5703920Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5704150Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5704192Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5704265Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5704302Z unimplemented [] 2025-12-04T09:45:06.5704358Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5704400Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5704439Z graph_break [] 2025-12-04T09:45:06.5704515Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5704743Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5704784Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5704857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5704894Z unimplemented [] 2025-12-04T09:45:06.5704950Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5704992Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5705028Z graph_break [] 2025-12-04T09:45:06.5705100Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5705330Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5705395Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5705468Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5705505Z unimplemented [] 2025-12-04T09:45:06.5705561Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5705603Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5705640Z graph_break [] 2025-12-04T09:45:06.5705712Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5705941Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5705983Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5706056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5706092Z unimplemented [] 2025-12-04T09:45:06.5706149Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5706193Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5706230Z graph_break [] 2025-12-04T09:45:06.5706302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5706528Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5706570Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5706643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5706679Z unimplemented [] 2025-12-04T09:45:06.5706735Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5706815Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5706852Z graph_break [] 2025-12-04T09:45:06.5706924Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5707182Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5707229Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5707303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5707339Z unimplemented [] 2025-12-04T09:45:06.5707395Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5707437Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5707473Z graph_break [] 2025-12-04T09:45:06.5707545Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5707773Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5707815Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5707890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5707928Z unimplemented [] 2025-12-04T09:45:06.5707984Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5708026Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5708062Z graph_break [] 2025-12-04T09:45:06.5708134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5708361Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5708402Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5708476Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5708513Z unimplemented [] 2025-12-04T09:45:06.5708569Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5708611Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5708651Z graph_break [] 2025-12-04T09:45:06.5708723Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5708980Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5709021Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5709127Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5709172Z Traceback (most recent call last): 2025-12-04T09:45:06.5709293Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5709331Z return value(self) 2025-12-04T09:45:06.5709480Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5709531Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5709673Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5709733Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5709892Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5709966Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5710020Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5710022Z 2025-12-04T09:45:06.5710065Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5710166Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5710268Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5710270Z 2025-12-04T09:45:06.5710343Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5710546Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5710550Z 2025-12-04T09:45:06.5710636Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5710711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5710749Z unimplemented [] 2025-12-04T09:45:06.5710806Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5710849Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5710886Z graph_break [] 2025-12-04T09:45:06.5710959Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5711190Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5711232Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5711307Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5711345Z unimplemented [] 2025-12-04T09:45:06.5711402Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5711444Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5711481Z graph_break [] 2025-12-04T09:45:06.5711553Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5711783Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5711824Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5711898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5711935Z unimplemented [] 2025-12-04T09:45:06.5711991Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5712033Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5712070Z graph_break [] 2025-12-04T09:45:06.5712145Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5712397Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5712439Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5712512Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5712550Z unimplemented [] 2025-12-04T09:45:06.5712605Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5712647Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5712685Z graph_break [] 2025-12-04T09:45:06.5712756Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5712985Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5713025Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5713100Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5713137Z unimplemented [] 2025-12-04T09:45:06.5713193Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5713234Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5713271Z graph_break [] 2025-12-04T09:45:06.5713343Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5713568Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5713609Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5713682Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5713719Z unimplemented [] 2025-12-04T09:45:06.5713800Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5713843Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5713882Z graph_break [] 2025-12-04T09:45:06.5713954Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5714183Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5714224Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5714297Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5714334Z unimplemented [] 2025-12-04T09:45:06.5714389Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5714430Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5714468Z graph_break [] 2025-12-04T09:45:06.5714540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5714769Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5714812Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5714884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5714921Z unimplemented [] 2025-12-04T09:45:06.5714977Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5715019Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5715056Z graph_break [] 2025-12-04T09:45:06.5715128Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5715356Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5715402Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5718032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5718120Z unimplemented [] 2025-12-04T09:45:06.5718176Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5718219Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5718257Z graph_break [] 2025-12-04T09:45:06.5718329Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5718560Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5718602Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5718676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5718713Z unimplemented [] 2025-12-04T09:45:06.5718774Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5718816Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5718853Z graph_break [] 2025-12-04T09:45:06.5718927Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5719157Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5719199Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5719270Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5719308Z unimplemented [] 2025-12-04T09:45:06.5719363Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5719406Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5719442Z graph_break [] 2025-12-04T09:45:06.5719512Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5719770Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5719813Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5719887Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5719925Z unimplemented [] 2025-12-04T09:45:06.5719979Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5720021Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5720056Z graph_break [] 2025-12-04T09:45:06.5720129Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5720355Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5720398Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5720468Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5720505Z unimplemented [] 2025-12-04T09:45:06.5720561Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5720604Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5720640Z graph_break [] 2025-12-04T09:45:06.5720712Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5720943Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5720985Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5721056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5721093Z unimplemented [] 2025-12-04T09:45:06.5721146Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5721189Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5721224Z graph_break [] 2025-12-04T09:45:06.5721296Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5721523Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5721590Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5721661Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5721698Z unimplemented [] 2025-12-04T09:45:06.5721752Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5721794Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5721830Z graph_break [] 2025-12-04T09:45:06.5721903Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5722129Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5722171Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5722244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5722281Z unimplemented [] 2025-12-04T09:45:06.5722337Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5722380Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5722415Z graph_break [] 2025-12-04T09:45:06.5722488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5722714Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5722755Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5722829Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5722866Z unimplemented [] 2025-12-04T09:45:06.5722921Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5722965Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5723001Z graph_break [] 2025-12-04T09:45:06.5723094Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5723324Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5723366Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5723438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5723475Z unimplemented [] 2025-12-04T09:45:06.5723528Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5723571Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5723607Z graph_break [] 2025-12-04T09:45:06.5723682Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5723911Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5723955Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5724027Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5724065Z unimplemented [] 2025-12-04T09:45:06.5724118Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5724162Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5724197Z graph_break [] 2025-12-04T09:45:06.5724269Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5724495Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5724537Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5724610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5724650Z unimplemented [] 2025-12-04T09:45:06.5724706Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5724752Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5724820Z graph_break [] 2025-12-04T09:45:06.5724896Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5725129Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5725172Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5725244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5725281Z unimplemented [] 2025-12-04T09:45:06.5725335Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5725378Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5725413Z graph_break [] 2025-12-04T09:45:06.5725486Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5725713Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5725758Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5725829Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5725868Z unimplemented [] 2025-12-04T09:45:06.5725922Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5725964Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5726000Z graph_break [] 2025-12-04T09:45:06.5726073Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5726300Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5726342Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5726434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5726473Z unimplemented [] 2025-12-04T09:45:06.5726529Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5726572Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5726608Z graph_break [] 2025-12-04T09:45:06.5726681Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5726944Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5726987Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5727058Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5727095Z unimplemented [] 2025-12-04T09:45:06.5727149Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5727192Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5727227Z graph_break [] 2025-12-04T09:45:06.5727301Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5727527Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5727570Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5727642Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5727679Z unimplemented [] 2025-12-04T09:45:06.5727737Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5727779Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5727814Z graph_break [] 2025-12-04T09:45:06.5727886Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5728112Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5728154Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5728226Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5728291Z unimplemented [] 2025-12-04T09:45:06.5728345Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5728387Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5728423Z graph_break [] 2025-12-04T09:45:06.5728495Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5728722Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5728762Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5728835Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5728871Z unimplemented [] 2025-12-04T09:45:06.5728925Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5728968Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5729006Z graph_break [] 2025-12-04T09:45:06.5729080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5729308Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5729349Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5729421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5729458Z unimplemented [] 2025-12-04T09:45:06.5729513Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5729554Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5729590Z graph_break [] 2025-12-04T09:45:06.5729661Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5729917Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5729961Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5730034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5730070Z unimplemented [] 2025-12-04T09:45:06.5730125Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5730167Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5730203Z graph_break [] 2025-12-04T09:45:06.5730274Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5730501Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5730542Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5730614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5730652Z unimplemented [] 2025-12-04T09:45:06.5730708Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5730751Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5730789Z graph_break [] 2025-12-04T09:45:06.5730861Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5731088Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5731129Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5731202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5731239Z unimplemented [] 2025-12-04T09:45:06.5731295Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5731336Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5731373Z graph_break [] 2025-12-04T09:45:06.5731447Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5731676Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5731738Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5731810Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5731847Z unimplemented [] 2025-12-04T09:45:06.5731902Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5731945Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5731983Z graph_break [] 2025-12-04T09:45:06.5732054Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5732281Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5732324Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5732398Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5732436Z unimplemented [] 2025-12-04T09:45:06.5732492Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5732534Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5732570Z graph_break [] 2025-12-04T09:45:06.5732642Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5732868Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5732912Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5732988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5733025Z unimplemented [] 2025-12-04T09:45:06.5733081Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5733141Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5733179Z graph_break [] 2025-12-04T09:45:06.5733255Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5733485Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5733527Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5733632Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5733679Z Traceback (most recent call last): 2025-12-04T09:45:06.5733802Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5733841Z return value(self) 2025-12-04T09:45:06.5733993Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5734046Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5734188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5734251Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5734411Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5734485Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5734539Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5734541Z 2025-12-04T09:45:06.5734588Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5734690Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5734794Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5734797Z 2025-12-04T09:45:06.5734870Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5735055Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5735077Z 2025-12-04T09:45:06.5735165Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5735239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5735276Z unimplemented [] 2025-12-04T09:45:06.5735333Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5735376Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5735414Z graph_break [] 2025-12-04T09:45:06.5735487Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5735721Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5735763Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5735838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5735877Z unimplemented [] 2025-12-04T09:45:06.5735934Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5735976Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5736014Z graph_break [] 2025-12-04T09:45:06.5736086Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5736313Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5736355Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5736428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5736465Z unimplemented [] 2025-12-04T09:45:06.5736521Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5736563Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5736621Z graph_break [] 2025-12-04T09:45:06.5736693Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5736961Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5737004Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5737077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5737117Z unimplemented [] 2025-12-04T09:45:06.5737174Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5737216Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5737255Z graph_break [] 2025-12-04T09:45:06.5737326Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5737558Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5737601Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5737676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5737713Z unimplemented [] 2025-12-04T09:45:06.5737770Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5737813Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5737853Z graph_break [] 2025-12-04T09:45:06.5737924Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5738156Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5738198Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5738273Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5738310Z unimplemented [] 2025-12-04T09:45:06.5738372Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5738448Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5738485Z graph_break [] 2025-12-04T09:45:06.5738558Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5738785Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5738827Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5738901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5738938Z unimplemented [] 2025-12-04T09:45:06.5738999Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5739041Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5739082Z graph_break [] 2025-12-04T09:45:06.5739155Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5739383Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5739425Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5739502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5739540Z unimplemented [] 2025-12-04T09:45:06.5739599Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5739643Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5739683Z graph_break [] 2025-12-04T09:45:06.5739756Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5739982Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5740024Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5740124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5740167Z unimplemented [] 2025-12-04T09:45:06.5740221Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5740266Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5740302Z graph_break [] 2025-12-04T09:45:06.5740376Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5740603Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5740650Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5740723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5740762Z unimplemented [] 2025-12-04T09:45:06.5740818Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5740862Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5740900Z graph_break [] 2025-12-04T09:45:06.5740976Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5741204Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5741250Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5741321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5741359Z unimplemented [] 2025-12-04T09:45:06.5741413Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5741457Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5741493Z graph_break [] 2025-12-04T09:45:06.5741567Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5741797Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5741861Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5741934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5741973Z unimplemented [] 2025-12-04T09:45:06.5742029Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5742073Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5742109Z graph_break [] 2025-12-04T09:45:06.5742182Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5742409Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5742453Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5742526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5742567Z unimplemented [] 2025-12-04T09:45:06.5742624Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5742669Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5742706Z graph_break [] 2025-12-04T09:45:06.5742781Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5743009Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5743051Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5743126Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5743163Z unimplemented [] 2025-12-04T09:45:06.5743223Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5743266Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5743304Z graph_break [] 2025-12-04T09:45:06.5743377Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5743625Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5743670Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5743742Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5743782Z unimplemented [] 2025-12-04T09:45:06.5743837Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5743879Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5743914Z graph_break [] 2025-12-04T09:45:06.5743994Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5744219Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5744260Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5744333Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5744373Z unimplemented [] 2025-12-04T09:45:06.5744428Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5744472Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5744508Z graph_break [] 2025-12-04T09:45:06.5744582Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5744807Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5744852Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5744925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5744963Z unimplemented [] 2025-12-04T09:45:06.5745018Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5745061Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5745097Z graph_break [] 2025-12-04T09:45:06.5745173Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5745422Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5745464Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5745535Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5745574Z unimplemented [] 2025-12-04T09:45:06.5745628Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5745670Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5745706Z graph_break [] 2025-12-04T09:45:06.5745781Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5746015Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5746059Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5746133Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5746173Z unimplemented [] 2025-12-04T09:45:06.5746226Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5746270Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5746306Z graph_break [] 2025-12-04T09:45:06.5746381Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5746606Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5746650Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5746721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5746789Z unimplemented [] 2025-12-04T09:45:06.5746871Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5746916Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5746953Z graph_break [] 2025-12-04T09:45:06.5747026Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5747252Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5747295Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5747366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5747406Z unimplemented [] 2025-12-04T09:45:06.5747461Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5747503Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5747539Z graph_break [] 2025-12-04T09:45:06.5747614Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5747841Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5747887Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5747958Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5747996Z unimplemented [] 2025-12-04T09:45:06.5748051Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5748097Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5748132Z graph_break [] 2025-12-04T09:45:06.5748205Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5748433Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5748476Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5748550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5748613Z unimplemented [] 2025-12-04T09:45:06.5748667Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5748712Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5748748Z graph_break [] 2025-12-04T09:45:06.5748822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5749047Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5749090Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5749163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5749200Z unimplemented [] 2025-12-04T09:45:06.5749255Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5749296Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5749332Z graph_break [] 2025-12-04T09:45:06.5749406Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5749634Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5749675Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5749747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5749783Z unimplemented [] 2025-12-04T09:45:06.5749839Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5749881Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5749918Z graph_break [] 2025-12-04T09:45:06.5749989Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5750242Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5750284Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5750362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5750400Z unimplemented [] 2025-12-04T09:45:06.5750455Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5750498Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5750534Z graph_break [] 2025-12-04T09:45:06.5750607Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5750836Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5750877Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5750951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5750988Z unimplemented [] 2025-12-04T09:45:06.5751044Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5751086Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5751124Z graph_break [] 2025-12-04T09:45:06.5751197Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5751426Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5751467Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5751540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5751576Z unimplemented [] 2025-12-04T09:45:06.5751631Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5751673Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5751710Z graph_break [] 2025-12-04T09:45:06.5751782Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5752011Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5752078Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5752151Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5752188Z unimplemented [] 2025-12-04T09:45:06.5752243Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5752284Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5752319Z graph_break [] 2025-12-04T09:45:06.5752392Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5752624Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5752665Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5752739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5752775Z unimplemented [] 2025-12-04T09:45:06.5752831Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5752873Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5752909Z graph_break [] 2025-12-04T09:45:06.5752980Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5753207Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5753252Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5753327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5753364Z unimplemented [] 2025-12-04T09:45:06.5753419Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5753461Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5753497Z graph_break [] 2025-12-04T09:45:06.5753587Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5753816Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5753857Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5753931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5753967Z unimplemented [] 2025-12-04T09:45:06.5754024Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5754065Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5754101Z graph_break [] 2025-12-04T09:45:06.5754173Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5754400Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5754441Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5754516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5754552Z unimplemented [] 2025-12-04T09:45:06.5754607Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5754648Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5754685Z graph_break [] 2025-12-04T09:45:06.5754756Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5754983Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5755024Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5755097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5755134Z unimplemented [] 2025-12-04T09:45:06.5755189Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5757111Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5757182Z graph_break [] 2025-12-04T09:45:06.5757254Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5757480Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5757521Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5757594Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5757631Z unimplemented [] 2025-12-04T09:45:06.5757686Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5757727Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5757764Z graph_break [] 2025-12-04T09:45:06.5757841Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5758071Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5758113Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5758218Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5758263Z Traceback (most recent call last): 2025-12-04T09:45:06.5758385Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5758423Z return value(self) 2025-12-04T09:45:06.5758573Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5758624Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5758762Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5758821Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5759005Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5759080Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5759133Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5759135Z 2025-12-04T09:45:06.5759178Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5759281Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5759383Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5759385Z 2025-12-04T09:45:06.5759458Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5759641Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5759644Z 2025-12-04T09:45:06.5759731Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5759805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5759844Z unimplemented [] 2025-12-04T09:45:06.5759901Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5759944Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5759981Z graph_break [] 2025-12-04T09:45:06.5760056Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5760291Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5760334Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5760406Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5760443Z unimplemented [] 2025-12-04T09:45:06.5760499Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5760542Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5760579Z graph_break [] 2025-12-04T09:45:06.5760674Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5760901Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5760943Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5761016Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5761053Z unimplemented [] 2025-12-04T09:45:06.5761108Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5761150Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5761186Z graph_break [] 2025-12-04T09:45:06.5761257Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5761486Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5761529Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5761601Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5761638Z unimplemented [] 2025-12-04T09:45:06.5761693Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5761734Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5761770Z graph_break [] 2025-12-04T09:45:06.5761842Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5762073Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5762114Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5762188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5762243Z unimplemented [] 2025-12-04T09:45:06.5762300Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5762343Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5762379Z graph_break [] 2025-12-04T09:45:06.5762450Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5762682Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5762723Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5762794Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5762831Z unimplemented [] 2025-12-04T09:45:06.5762886Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5762927Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5762963Z graph_break [] 2025-12-04T09:45:06.5763036Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5763263Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5763306Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5763378Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5763415Z unimplemented [] 2025-12-04T09:45:06.5763469Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5763511Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5763547Z graph_break [] 2025-12-04T09:45:06.5763618Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5763845Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5763885Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5763960Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5764015Z unimplemented [] 2025-12-04T09:45:06.5764071Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5764113Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5764149Z graph_break [] 2025-12-04T09:45:06.5764223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5764450Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5764492Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5764563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5764601Z unimplemented [] 2025-12-04T09:45:06.5764655Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5764697Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5764734Z graph_break [] 2025-12-04T09:45:06.5764807Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5765034Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5765075Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5765146Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5765184Z unimplemented [] 2025-12-04T09:45:06.5765238Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5765279Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5765314Z graph_break [] 2025-12-04T09:45:06.5765387Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5765641Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5765686Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5765757Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5765795Z unimplemented [] 2025-12-04T09:45:06.5765849Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5765891Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5765926Z graph_break [] 2025-12-04T09:45:06.5765999Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5766227Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5766271Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5766342Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5766381Z unimplemented [] 2025-12-04T09:45:06.5766436Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5766480Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5766515Z graph_break [] 2025-12-04T09:45:06.5766587Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5766845Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5766887Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5766958Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5766996Z unimplemented [] 2025-12-04T09:45:06.5767049Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5767092Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5767127Z graph_break [] 2025-12-04T09:45:06.5767198Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5767425Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5767495Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5767566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5767604Z unimplemented [] 2025-12-04T09:45:06.5767657Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5767700Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5767735Z graph_break [] 2025-12-04T09:45:06.5767807Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5768033Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5768074Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5768147Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5768186Z unimplemented [] 2025-12-04T09:45:06.5768240Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5768283Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5768318Z graph_break [] 2025-12-04T09:45:06.5768390Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5768616Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5768657Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5768729Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5768766Z unimplemented [] 2025-12-04T09:45:06.5768819Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5768861Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5768925Z graph_break [] 2025-12-04T09:45:06.5768998Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5769226Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5769267Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5769338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5769376Z unimplemented [] 2025-12-04T09:45:06.5769429Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5769472Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5769507Z graph_break [] 2025-12-04T09:45:06.5769578Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5769806Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5769850Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5769922Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5769959Z unimplemented [] 2025-12-04T09:45:06.5770013Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5770056Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5770091Z graph_break [] 2025-12-04T09:45:06.5770163Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5770394Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5770436Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5770507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5770545Z unimplemented [] 2025-12-04T09:45:06.5770600Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5770663Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5770698Z graph_break [] 2025-12-04T09:45:06.5770771Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5770998Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5771041Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5771112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5771149Z unimplemented [] 2025-12-04T09:45:06.5771204Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5771248Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5771284Z graph_break [] 2025-12-04T09:45:06.5771358Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5771585Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5771628Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5771700Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5771738Z unimplemented [] 2025-12-04T09:45:06.5771792Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5771835Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5771871Z graph_break [] 2025-12-04T09:45:06.5771944Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5772168Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5772211Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5772301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5772342Z unimplemented [] 2025-12-04T09:45:06.5772396Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5772440Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5772475Z graph_break [] 2025-12-04T09:45:06.5772548Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5772774Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5772818Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5772894Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5772930Z unimplemented [] 2025-12-04T09:45:06.5772985Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5773027Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5773064Z graph_break [] 2025-12-04T09:45:06.5773137Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5773364Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5773405Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5773477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5773514Z unimplemented [] 2025-12-04T09:45:06.5773569Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5773611Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5773647Z graph_break [] 2025-12-04T09:45:06.5773719Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5773946Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5774005Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5774077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5774113Z unimplemented [] 2025-12-04T09:45:06.5774168Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5774210Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5774245Z graph_break [] 2025-12-04T09:45:06.5774317Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5774542Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5774582Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5774654Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5774690Z unimplemented [] 2025-12-04T09:45:06.5774746Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5774789Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5774825Z graph_break [] 2025-12-04T09:45:06.5774902Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5775132Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5775174Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5775246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5775283Z unimplemented [] 2025-12-04T09:45:06.5775338Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5775379Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5775415Z graph_break [] 2025-12-04T09:45:06.5775487Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5775732Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5775776Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5775848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5775885Z unimplemented [] 2025-12-04T09:45:06.5775941Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5775982Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5776018Z graph_break [] 2025-12-04T09:45:06.5776089Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5776317Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5776357Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5776430Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5776469Z unimplemented [] 2025-12-04T09:45:06.5776524Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5776566Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5776602Z graph_break [] 2025-12-04T09:45:06.5776673Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5776938Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5776979Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5777051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5777087Z unimplemented [] 2025-12-04T09:45:06.5777142Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5777184Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5777221Z graph_break [] 2025-12-04T09:45:06.5777297Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5777557Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5777599Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5777671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5777707Z unimplemented [] 2025-12-04T09:45:06.5777762Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5777804Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5777840Z graph_break [] 2025-12-04T09:45:06.5777911Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5778141Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5778182Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5778257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5778293Z unimplemented [] 2025-12-04T09:45:06.5778349Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5778391Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5778429Z graph_break [] 2025-12-04T09:45:06.5778500Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5778727Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5778769Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5778842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5778878Z unimplemented [] 2025-12-04T09:45:06.5778957Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5778999Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5779038Z graph_break [] 2025-12-04T09:45:06.5779110Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5779339Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5779380Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5779452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5779489Z unimplemented [] 2025-12-04T09:45:06.5779543Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5779585Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5779621Z graph_break [] 2025-12-04T09:45:06.5779691Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5779920Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5779962Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5780035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5780071Z unimplemented [] 2025-12-04T09:45:06.5780126Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5780168Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5780204Z graph_break [] 2025-12-04T09:45:06.5780275Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5780502Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5780543Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5780617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5780675Z unimplemented [] 2025-12-04T09:45:06.5780730Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5780772Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5780808Z graph_break [] 2025-12-04T09:45:06.5780880Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5781107Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5781150Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5781253Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5781299Z Traceback (most recent call last): 2025-12-04T09:45:06.5781418Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5781458Z return value(self) 2025-12-04T09:45:06.5781607Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5781661Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5781797Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5781858Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5782016Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5782089Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5782140Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5782142Z 2025-12-04T09:45:06.5782186Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5782286Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5782408Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5782411Z 2025-12-04T09:45:06.5782484Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5782667Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5782669Z 2025-12-04T09:45:06.5782755Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5782827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5782865Z unimplemented [] 2025-12-04T09:45:06.5782921Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5782964Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5783001Z graph_break [] 2025-12-04T09:45:06.5783073Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5783306Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5783349Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5783422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5783459Z unimplemented [] 2025-12-04T09:45:06.5783515Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5783557Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5783593Z graph_break [] 2025-12-04T09:45:06.5783665Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5783892Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5783934Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5784008Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5784044Z unimplemented [] 2025-12-04T09:45:06.5784120Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5784163Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5784202Z graph_break [] 2025-12-04T09:45:06.5784274Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5784502Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5784544Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5784618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5784655Z unimplemented [] 2025-12-04T09:45:06.5784711Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5784753Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5784791Z graph_break [] 2025-12-04T09:45:06.5784864Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5785097Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5785138Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5785212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5785250Z unimplemented [] 2025-12-04T09:45:06.5785307Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5785350Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5785387Z graph_break [] 2025-12-04T09:45:06.5785459Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5785688Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5785750Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5785825Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5785864Z unimplemented [] 2025-12-04T09:45:06.5785918Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5785960Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5785999Z graph_break [] 2025-12-04T09:45:06.5786070Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5786299Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5786343Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5786415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5786453Z unimplemented [] 2025-12-04T09:45:06.5786509Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5786555Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5786592Z graph_break [] 2025-12-04T09:45:06.5786666Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5786917Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5786960Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5787032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5787071Z unimplemented [] 2025-12-04T09:45:06.5787125Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5787170Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5787206Z graph_break [] 2025-12-04T09:45:06.5787279Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5787510Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5787583Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5787654Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5787692Z unimplemented [] 2025-12-04T09:45:06.5787748Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5787792Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5787827Z graph_break [] 2025-12-04T09:45:06.5787901Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5788127Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5788168Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5788239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5788281Z unimplemented [] 2025-12-04T09:45:06.5788337Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5788380Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5788417Z graph_break [] 2025-12-04T09:45:06.5788490Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5788714Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5788757Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5788830Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5788866Z unimplemented [] 2025-12-04T09:45:06.5788922Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5788968Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5789004Z graph_break [] 2025-12-04T09:45:06.5789103Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5789331Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5789375Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5789447Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5789488Z unimplemented [] 2025-12-04T09:45:06.5789545Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5789587Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5789621Z graph_break [] 2025-12-04T09:45:06.5789697Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5789923Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5789967Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5790040Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5790079Z unimplemented [] 2025-12-04T09:45:06.5790133Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5790178Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5790214Z graph_break [] 2025-12-04T09:45:06.5790289Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5790515Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5790558Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5790628Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5790667Z unimplemented [] 2025-12-04T09:45:06.5790724Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5790769Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5790804Z graph_break [] 2025-12-04T09:45:06.5790902Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5791128Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5791174Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5791246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5791283Z unimplemented [] 2025-12-04T09:45:06.5791337Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5791380Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5791416Z graph_break [] 2025-12-04T09:45:06.5791489Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5791717Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5791763Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5791834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5791874Z unimplemented [] 2025-12-04T09:45:06.5791927Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5791970Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5792007Z graph_break [] 2025-12-04T09:45:06.5792080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5792304Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5792348Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5792421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5792477Z unimplemented [] 2025-12-04T09:45:06.5792532Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5792579Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5792617Z graph_break [] 2025-12-04T09:45:06.5792692Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5792918Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5792961Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5793034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5793074Z unimplemented [] 2025-12-04T09:45:06.5793128Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5793174Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5793209Z graph_break [] 2025-12-04T09:45:06.5793285Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5793511Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5793556Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5793628Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5793665Z unimplemented [] 2025-12-04T09:45:06.5793721Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5793766Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5793801Z graph_break [] 2025-12-04T09:45:06.5793874Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5794100Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5794144Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5794218Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5794277Z unimplemented [] 2025-12-04T09:45:06.5794332Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5794378Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5794413Z graph_break [] 2025-12-04T09:45:06.5794485Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5794712Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5794754Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5794828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5794866Z unimplemented [] 2025-12-04T09:45:06.5794921Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5794964Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5795001Z graph_break [] 2025-12-04T09:45:06.5795076Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5795306Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5795347Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5795421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5795458Z unimplemented [] 2025-12-04T09:45:06.5795515Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5795558Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5795596Z graph_break [] 2025-12-04T09:45:06.5795668Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5795917Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5795961Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5796034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5796072Z unimplemented [] 2025-12-04T09:45:06.5796129Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5796172Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5796210Z graph_break [] 2025-12-04T09:45:06.5796282Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5796511Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5796551Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5796624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5796661Z unimplemented [] 2025-12-04T09:45:06.5796720Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5796798Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5796835Z graph_break [] 2025-12-04T09:45:06.5796907Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5797133Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5797174Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5797247Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5797284Z unimplemented [] 2025-12-04T09:45:06.5797340Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5797381Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5797417Z graph_break [] 2025-12-04T09:45:06.5797488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5797716Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5797785Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5797858Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5797894Z unimplemented [] 2025-12-04T09:45:06.5797950Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5797992Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5798029Z graph_break [] 2025-12-04T09:45:06.5798099Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5798326Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5798367Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5798440Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5798478Z unimplemented [] 2025-12-04T09:45:06.5798534Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5798575Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5798611Z graph_break [] 2025-12-04T09:45:06.5798683Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5798909Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5798950Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5799023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5799059Z unimplemented [] 2025-12-04T09:45:06.5799114Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5799156Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5799219Z graph_break [] 2025-12-04T09:45:06.5799291Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5799519Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5799561Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5799633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5799669Z unimplemented [] 2025-12-04T09:45:06.5799724Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5799765Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5799801Z graph_break [] 2025-12-04T09:45:06.5799872Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5800102Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5800144Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5800216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5800253Z unimplemented [] 2025-12-04T09:45:06.5800308Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5800349Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5800386Z graph_break [] 2025-12-04T09:45:06.5800457Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5800683Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5800725Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5800798Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5800834Z unimplemented [] 2025-12-04T09:45:06.5800891Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5800953Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5800990Z graph_break [] 2025-12-04T09:45:06.5801062Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5801287Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5801328Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5801401Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5801437Z unimplemented [] 2025-12-04T09:45:06.5801492Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5801534Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5801571Z graph_break [] 2025-12-04T09:45:06.5801643Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5801870Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5801912Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5801984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5802021Z unimplemented [] 2025-12-04T09:45:06.5802076Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5802118Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5802154Z graph_break [] 2025-12-04T09:45:06.5802225Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5802451Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5802493Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5802589Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5802628Z unimplemented [] 2025-12-04T09:45:06.5802683Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5802725Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5802761Z graph_break [] 2025-12-04T09:45:06.5802832Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5803059Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5803101Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5803172Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5803210Z unimplemented [] 2025-12-04T09:45:06.5803264Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5803307Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5803345Z graph_break [] 2025-12-04T09:45:06.5803418Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5803644Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5803686Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5803758Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5803795Z unimplemented [] 2025-12-04T09:45:06.5803848Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5803891Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5803927Z graph_break [] 2025-12-04T09:45:06.5803999Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5804226Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5804288Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5804360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5804397Z unimplemented [] 2025-12-04T09:45:06.5804452Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5804494Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5804529Z graph_break [] 2025-12-04T09:45:06.5804601Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5804827Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5804868Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5804971Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5805017Z Traceback (most recent call last): 2025-12-04T09:45:06.5805137Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5805177Z return value(self) 2025-12-04T09:45:06.5805324Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5805376Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5805511Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5805570Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5805728Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5805800Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5805851Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5805853Z 2025-12-04T09:45:06.5805897Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5806017Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5806121Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5806123Z 2025-12-04T09:45:06.5806195Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5806378Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5806381Z 2025-12-04T09:45:06.5806466Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5806539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5806577Z unimplemented [] 2025-12-04T09:45:06.5806633Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5806675Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5806712Z graph_break [] 2025-12-04T09:45:06.5806826Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5807057Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5807098Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5807171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5807208Z unimplemented [] 2025-12-04T09:45:06.5807264Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5807305Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5807342Z graph_break [] 2025-12-04T09:45:06.5807413Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5807642Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5807684Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5807789Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5807827Z unimplemented [] 2025-12-04T09:45:06.5807881Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5807924Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5807959Z graph_break [] 2025-12-04T09:45:06.5808032Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5808257Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5808299Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5808371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5808408Z unimplemented [] 2025-12-04T09:45:06.5808464Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5808507Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5808544Z graph_break [] 2025-12-04T09:45:06.5808616Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5808841Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5808883Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5808955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5808993Z unimplemented [] 2025-12-04T09:45:06.5809046Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5809089Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5809124Z graph_break [] 2025-12-04T09:45:06.5809196Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5809450Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5809495Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5809566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5809603Z unimplemented [] 2025-12-04T09:45:06.5809657Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5809701Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5809736Z graph_break [] 2025-12-04T09:45:06.5809808Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5810033Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5810075Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5810149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5810187Z unimplemented [] 2025-12-04T09:45:06.5810243Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5810286Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5810320Z graph_break [] 2025-12-04T09:45:06.5810394Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5810620Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5810662Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5810733Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5810770Z unimplemented [] 2025-12-04T09:45:06.5810824Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5810868Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5810902Z graph_break [] 2025-12-04T09:45:06.5810977Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5811226Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5811268Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5811339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5811377Z unimplemented [] 2025-12-04T09:45:06.5811431Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5811474Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5811509Z graph_break [] 2025-12-04T09:45:06.5811581Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5811805Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5811849Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5811922Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5811959Z unimplemented [] 2025-12-04T09:45:06.5812013Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5812055Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5812090Z graph_break [] 2025-12-04T09:45:06.5812162Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5812387Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5812429Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5812500Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5812537Z unimplemented [] 2025-12-04T09:45:06.5812591Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5812654Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5812691Z graph_break [] 2025-12-04T09:45:06.5812764Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5812991Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5813033Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5813103Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5813140Z unimplemented [] 2025-12-04T09:45:06.5813194Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5813237Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5813272Z graph_break [] 2025-12-04T09:45:06.5813344Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5813571Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5813615Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5813686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5813723Z unimplemented [] 2025-12-04T09:45:06.5813777Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5813820Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5813855Z graph_break [] 2025-12-04T09:45:06.5813929Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5814153Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5814195Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5814268Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5814306Z unimplemented [] 2025-12-04T09:45:06.5814379Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5814422Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5814457Z graph_break [] 2025-12-04T09:45:06.5814532Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5814758Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5814800Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5814871Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5814909Z unimplemented [] 2025-12-04T09:45:06.5814963Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5815006Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5815041Z graph_break [] 2025-12-04T09:45:06.5815115Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5815342Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5815384Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5815455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5815492Z unimplemented [] 2025-12-04T09:45:06.5815545Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5815588Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5815623Z graph_break [] 2025-12-04T09:45:06.5815694Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5815920Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5815987Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5816062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5816099Z unimplemented [] 2025-12-04T09:45:06.5816153Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5816195Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5816230Z graph_break [] 2025-12-04T09:45:06.5816302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5816528Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5816569Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5816642Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5816678Z unimplemented [] 2025-12-04T09:45:06.5816733Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5816817Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5816854Z graph_break [] 2025-12-04T09:45:06.5816927Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5817154Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5817196Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5817268Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5817304Z unimplemented [] 2025-12-04T09:45:06.5817359Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5817401Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5817437Z graph_break [] 2025-12-04T09:45:06.5817508Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5817736Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5817806Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5817879Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5817915Z unimplemented [] 2025-12-04T09:45:06.5817970Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5818012Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5818047Z graph_break [] 2025-12-04T09:45:06.5818119Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5818344Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5818385Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5818457Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5818495Z unimplemented [] 2025-12-04T09:45:06.5818551Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5818593Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5818629Z graph_break [] 2025-12-04T09:45:06.5818700Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5818929Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5818970Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5819042Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5819078Z unimplemented [] 2025-12-04T09:45:06.5819133Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5819174Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5819210Z graph_break [] 2025-12-04T09:45:06.5819306Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5819535Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5819576Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5819648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5819685Z unimplemented [] 2025-12-04T09:45:06.5819740Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5819782Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5819817Z graph_break [] 2025-12-04T09:45:06.5819888Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5820115Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5820157Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5820229Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5820266Z unimplemented [] 2025-12-04T09:45:06.5820320Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5820362Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5820398Z graph_break [] 2025-12-04T09:45:06.5820469Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5820703Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5820744Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5820816Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5820852Z unimplemented [] 2025-12-04T09:45:06.5820907Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5820950Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5820986Z graph_break [] 2025-12-04T09:45:06.5821077Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5821306Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5821347Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5821419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5821456Z unimplemented [] 2025-12-04T09:45:06.5821511Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5821553Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5821589Z graph_break [] 2025-12-04T09:45:06.5821661Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5821889Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5821932Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5822004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5822040Z unimplemented [] 2025-12-04T09:45:06.5822095Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5822137Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5822174Z graph_break [] 2025-12-04T09:45:06.5822245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5822474Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5823961Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5824041Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5824107Z unimplemented [] 2025-12-04T09:45:06.5824165Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5824209Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5824246Z graph_break [] 2025-12-04T09:45:06.5824318Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5824547Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5824589Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5824662Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5824700Z unimplemented [] 2025-12-04T09:45:06.5824759Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5824801Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5824837Z graph_break [] 2025-12-04T09:45:06.5824910Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5825139Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5825181Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5825254Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5825290Z unimplemented [] 2025-12-04T09:45:06.5825347Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5825389Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5825425Z graph_break [] 2025-12-04T09:45:06.5825497Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5825725Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5825768Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5825842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5825898Z unimplemented [] 2025-12-04T09:45:06.5825953Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5825995Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5826031Z graph_break [] 2025-12-04T09:45:06.5826102Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5826330Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5826373Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5826445Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5826482Z unimplemented [] 2025-12-04T09:45:06.5826536Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5826579Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5826615Z graph_break [] 2025-12-04T09:45:06.5826688Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5826953Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5826995Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5827066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5827103Z unimplemented [] 2025-12-04T09:45:06.5827157Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5827200Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5827235Z graph_break [] 2025-12-04T09:45:06.5827307Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5827565Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5827609Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5827680Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5827718Z unimplemented [] 2025-12-04T09:45:06.5827772Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5827815Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5827850Z graph_break [] 2025-12-04T09:45:06.5827922Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5828147Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5828189Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5828261Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5828301Z unimplemented [] 2025-12-04T09:45:06.5828355Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5828399Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5828434Z graph_break [] 2025-12-04T09:45:06.5828506Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5828732Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5828774Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5828845Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5828883Z unimplemented [] 2025-12-04T09:45:06.5828937Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5828980Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5829014Z graph_break [] 2025-12-04T09:45:06.5829088Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5829313Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5829386Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5829458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5829495Z unimplemented [] 2025-12-04T09:45:06.5829549Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5829591Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5829626Z graph_break [] 2025-12-04T09:45:06.5829701Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5829925Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5829967Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5830040Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5830079Z unimplemented [] 2025-12-04T09:45:06.5830133Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5830176Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5830211Z graph_break [] 2025-12-04T09:45:06.5830285Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5830515Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5830556Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5830660Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5830705Z Traceback (most recent call last): 2025-12-04T09:45:06.5830826Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5830889Z return value(self) 2025-12-04T09:45:06.5831040Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5831093Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5831231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5831293Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5831452Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5831527Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5831579Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5831581Z 2025-12-04T09:45:06.5831625Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5831729Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5831833Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5831836Z 2025-12-04T09:45:06.5831910Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5832094Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5832097Z 2025-12-04T09:45:06.5832184Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5832257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5832295Z unimplemented [] 2025-12-04T09:45:06.5832351Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5832394Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5832430Z graph_break [] 2025-12-04T09:45:06.5832504Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5832736Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5832800Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5832871Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5832909Z unimplemented [] 2025-12-04T09:45:06.5832963Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5833006Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5833041Z graph_break [] 2025-12-04T09:45:06.5833113Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5833341Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5833383Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5833456Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5833495Z unimplemented [] 2025-12-04T09:45:06.5833550Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5833592Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5833628Z graph_break [] 2025-12-04T09:45:06.5833700Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5833925Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5833968Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5834039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5834076Z unimplemented [] 2025-12-04T09:45:06.5834131Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5834174Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5834209Z graph_break [] 2025-12-04T09:45:06.5834300Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5834528Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5834570Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5834643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5834680Z unimplemented [] 2025-12-04T09:45:06.5834734Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5834776Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5834811Z graph_break [] 2025-12-04T09:45:06.5834883Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5835111Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5835152Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5835225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5835262Z unimplemented [] 2025-12-04T09:45:06.5835316Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5835358Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5835393Z graph_break [] 2025-12-04T09:45:06.5835464Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5835688Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5835730Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5835801Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5835838Z unimplemented [] 2025-12-04T09:45:06.5835893Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5835936Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5835991Z graph_break [] 2025-12-04T09:45:06.5836064Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5836289Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5836331Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5836403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5836440Z unimplemented [] 2025-12-04T09:45:06.5836493Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5836536Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5836571Z graph_break [] 2025-12-04T09:45:06.5836644Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5836900Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5836944Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5837016Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5837054Z unimplemented [] 2025-12-04T09:45:06.5837108Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5837151Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5837186Z graph_break [] 2025-12-04T09:45:06.5837259Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5837489Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5837532Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5837630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5837670Z unimplemented [] 2025-12-04T09:45:06.5837724Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5837769Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5837805Z graph_break [] 2025-12-04T09:45:06.5837878Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5838103Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5838145Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5838216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5838256Z unimplemented [] 2025-12-04T09:45:06.5838310Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5838352Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5838387Z graph_break [] 2025-12-04T09:45:06.5838461Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5838689Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5838730Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5838802Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5838841Z unimplemented [] 2025-12-04T09:45:06.5838894Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5838937Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5838972Z graph_break [] 2025-12-04T09:45:06.5839045Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5839274Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5839317Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5839419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5839456Z unimplemented [] 2025-12-04T09:45:06.5839510Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5839553Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5839589Z graph_break [] 2025-12-04T09:45:06.5839662Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5839888Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5839930Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5840003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5840040Z unimplemented [] 2025-12-04T09:45:06.5840098Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5840141Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5840178Z graph_break [] 2025-12-04T09:45:06.5840250Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5840476Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5840518Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5840590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5840627Z unimplemented [] 2025-12-04T09:45:06.5840682Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5840725Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5840761Z graph_break [] 2025-12-04T09:45:06.5840833Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5841081Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5841125Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5841197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5841233Z unimplemented [] 2025-12-04T09:45:06.5841288Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5841329Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5841365Z graph_break [] 2025-12-04T09:45:06.5841438Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5841664Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5841705Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5841779Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5841815Z unimplemented [] 2025-12-04T09:45:06.5841872Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5841913Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5841950Z graph_break [] 2025-12-04T09:45:06.5842022Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5842247Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5842288Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5842360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5842397Z unimplemented [] 2025-12-04T09:45:06.5842451Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5842495Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5842532Z graph_break [] 2025-12-04T09:45:06.5842605Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5842868Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5842909Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5842983Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5843021Z unimplemented [] 2025-12-04T09:45:06.5843077Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5843120Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5843156Z graph_break [] 2025-12-04T09:45:06.5843229Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5843454Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5843499Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5843576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5843613Z unimplemented [] 2025-12-04T09:45:06.5843667Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5843711Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5843748Z graph_break [] 2025-12-04T09:45:06.5843820Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5844048Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5844089Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5844163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5844198Z unimplemented [] 2025-12-04T09:45:06.5844254Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5844317Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5844355Z graph_break [] 2025-12-04T09:45:06.5844427Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5844653Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5844694Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5844767Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5844805Z unimplemented [] 2025-12-04T09:45:06.5844862Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5844905Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5844942Z graph_break [] 2025-12-04T09:45:06.5845015Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5845245Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5845288Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5845366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5845402Z unimplemented [] 2025-12-04T09:45:06.5845463Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5845505Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5845542Z graph_break [] 2025-12-04T09:45:06.5845614Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5845841Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5845884Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5845958Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5845997Z unimplemented [] 2025-12-04T09:45:06.5846081Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5846125Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5846162Z graph_break [] 2025-12-04T09:45:06.5846235Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5846464Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5846508Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5846581Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5846618Z unimplemented [] 2025-12-04T09:45:06.5846673Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5846715Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5846785Z graph_break [] 2025-12-04T09:45:06.5846862Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5847096Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5847139Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5847210Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5847247Z unimplemented [] 2025-12-04T09:45:06.5847302Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5847345Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5847380Z graph_break [] 2025-12-04T09:45:06.5847453Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5847680Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5847753Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5847827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5847868Z unimplemented [] 2025-12-04T09:45:06.5847926Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5847968Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5848004Z graph_break [] 2025-12-04T09:45:06.5848076Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5848301Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5848346Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5848417Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5848457Z unimplemented [] 2025-12-04T09:45:06.5848511Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5848558Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5848598Z graph_break [] 2025-12-04T09:45:06.5848672Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5848901Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5848946Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5849019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5849058Z unimplemented [] 2025-12-04T09:45:06.5849112Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5849156Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5849191Z graph_break [] 2025-12-04T09:45:06.5849266Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5849497Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5849570Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5849643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5849681Z unimplemented [] 2025-12-04T09:45:06.5849735Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5849778Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5849813Z graph_break [] 2025-12-04T09:45:06.5849886Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5850111Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5850154Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5850225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5850265Z unimplemented [] 2025-12-04T09:45:06.5850321Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5850363Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5850398Z graph_break [] 2025-12-04T09:45:06.5850471Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5850697Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5850740Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5850811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5850848Z unimplemented [] 2025-12-04T09:45:06.5850902Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5850944Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5850979Z graph_break [] 2025-12-04T09:45:06.5851074Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5851302Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5851344Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5851415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5851453Z unimplemented [] 2025-12-04T09:45:06.5851506Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5851549Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5851584Z graph_break [] 2025-12-04T09:45:06.5851656Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5851881Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5851924Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5851996Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5852035Z unimplemented [] 2025-12-04T09:45:06.5852089Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5852132Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5852167Z graph_break [] 2025-12-04T09:45:06.5852239Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5852465Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5852507Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5852578Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5852615Z unimplemented [] 2025-12-04T09:45:06.5852669Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5852713Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5852748Z graph_break [] 2025-12-04T09:45:06.5852845Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5853071Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5853112Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5853185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5853222Z unimplemented [] 2025-12-04T09:45:06.5853275Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5853317Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5853352Z graph_break [] 2025-12-04T09:45:06.5853426Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5853652Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5853696Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5853768Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5853805Z unimplemented [] 2025-12-04T09:45:06.5853858Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5853901Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5853935Z graph_break [] 2025-12-04T09:45:06.5854007Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5854232Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5854274Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5854345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5854412Z unimplemented [] 2025-12-04T09:45:06.5854467Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5854511Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5854546Z graph_break [] 2025-12-04T09:45:06.5854619Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5854845Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5854887Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5854959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5854996Z unimplemented [] 2025-12-04T09:45:06.5855049Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5855092Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5855127Z graph_break [] 2025-12-04T09:45:06.5855201Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5855430Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5855473Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5855577Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5855623Z Traceback (most recent call last): 2025-12-04T09:45:06.5855743Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5855783Z return value(self) 2025-12-04T09:45:06.5855932Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5855984Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5856122Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5856183Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5856373Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5856445Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5856498Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5856501Z 2025-12-04T09:45:06.5856544Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5856647Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5856780Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5856782Z 2025-12-04T09:45:06.5856854Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5857037Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5857040Z 2025-12-04T09:45:06.5857128Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5857201Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5857239Z unimplemented [] 2025-12-04T09:45:06.5857295Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5857339Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5857375Z graph_break [] 2025-12-04T09:45:06.5857448Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5857678Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5857721Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5857792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5857830Z unimplemented [] 2025-12-04T09:45:06.5857912Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5857957Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5857993Z graph_break [] 2025-12-04T09:45:06.5858066Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5858293Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5858336Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5858407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5858445Z unimplemented [] 2025-12-04T09:45:06.5858498Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5858541Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5858575Z graph_break [] 2025-12-04T09:45:06.5858648Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5858875Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5858919Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5858991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5859028Z unimplemented [] 2025-12-04T09:45:06.5859083Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5859125Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5859160Z graph_break [] 2025-12-04T09:45:06.5859232Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5859457Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5859499Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5859572Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5859645Z unimplemented [] 2025-12-04T09:45:06.5859699Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5859741Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5859776Z graph_break [] 2025-12-04T09:45:06.5859848Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5860074Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5860116Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5860187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5860224Z unimplemented [] 2025-12-04T09:45:06.5860278Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5860321Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5860358Z graph_break [] 2025-12-04T09:45:06.5860430Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5860658Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5860700Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5860771Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5860809Z unimplemented [] 2025-12-04T09:45:06.5860862Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5860905Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5860939Z graph_break [] 2025-12-04T09:45:06.5861011Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5861259Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5861303Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5861375Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5861412Z unimplemented [] 2025-12-04T09:45:06.5861467Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5861509Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5861545Z graph_break [] 2025-12-04T09:45:06.5861617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5861845Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5861886Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5861957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5861994Z unimplemented [] 2025-12-04T09:45:06.5862050Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5862093Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5862128Z graph_break [] 2025-12-04T09:45:06.5862200Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5862426Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5862467Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5862539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5862576Z unimplemented [] 2025-12-04T09:45:06.5862630Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5862671Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5862708Z graph_break [] 2025-12-04T09:45:06.5862780Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5863013Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5863083Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5863155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5863191Z unimplemented [] 2025-12-04T09:45:06.5863246Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5863287Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5863322Z graph_break [] 2025-12-04T09:45:06.5863394Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5863619Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5863659Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5863733Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5863771Z unimplemented [] 2025-12-04T09:45:06.5863827Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5863868Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5863904Z graph_break [] 2025-12-04T09:45:06.5863976Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5864201Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5864242Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5864313Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5864350Z unimplemented [] 2025-12-04T09:45:06.5864405Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5864446Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5864481Z graph_break [] 2025-12-04T09:45:06.5864571Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5864799Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5864839Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5864912Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5864948Z unimplemented [] 2025-12-04T09:45:06.5865003Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5865044Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5865080Z graph_break [] 2025-12-04T09:45:06.5865151Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5865380Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5865421Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5865496Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5865532Z unimplemented [] 2025-12-04T09:45:06.5865586Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5865628Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5865663Z graph_break [] 2025-12-04T09:45:06.5865736Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5865962Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5866003Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5866075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5866111Z unimplemented [] 2025-12-04T09:45:06.5866167Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5866240Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5866275Z graph_break [] 2025-12-04T09:45:06.5866347Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5866572Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5866613Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5866685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5866722Z unimplemented [] 2025-12-04T09:45:06.5866809Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5866851Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5866886Z graph_break [] 2025-12-04T09:45:06.5866958Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5867187Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5867229Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5867302Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5867339Z unimplemented [] 2025-12-04T09:45:06.5867394Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5867436Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5867472Z graph_break [] 2025-12-04T09:45:06.5867543Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5867772Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5867813Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5867922Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5867964Z unimplemented [] 2025-12-04T09:45:06.5868019Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5868060Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5868096Z graph_break [] 2025-12-04T09:45:06.5868168Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5868399Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5868440Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5868513Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5868549Z unimplemented [] 2025-12-04T09:45:06.5868604Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5868646Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5868682Z graph_break [] 2025-12-04T09:45:06.5868755Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5868983Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5869024Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5869096Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5869133Z unimplemented [] 2025-12-04T09:45:06.5869188Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5869229Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5869265Z graph_break [] 2025-12-04T09:45:06.5869336Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5869565Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5869607Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5869710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5869748Z unimplemented [] 2025-12-04T09:45:06.5869802Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5869844Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5869879Z graph_break [] 2025-12-04T09:45:06.5869950Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5870176Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5870218Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5870289Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5870326Z unimplemented [] 2025-12-04T09:45:06.5870382Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5870424Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5870461Z graph_break [] 2025-12-04T09:45:06.5870533Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5870758Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5870800Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5870871Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5870908Z unimplemented [] 2025-12-04T09:45:06.5870962Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5871004Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5871039Z graph_break [] 2025-12-04T09:45:06.5871111Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5871357Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5871401Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5871472Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5871509Z unimplemented [] 2025-12-04T09:45:06.5871563Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5871605Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5871640Z graph_break [] 2025-12-04T09:45:06.5871712Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5871939Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5871981Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5872054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5872091Z unimplemented [] 2025-12-04T09:45:06.5872146Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5872190Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5872225Z graph_break [] 2025-12-04T09:45:06.5872298Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5872522Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5872564Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5872635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5872671Z unimplemented [] 2025-12-04T09:45:06.5872725Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5872767Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5872802Z graph_break [] 2025-12-04T09:45:06.5872876Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5873126Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5873168Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5873239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5873276Z unimplemented [] 2025-12-04T09:45:06.5873329Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5873371Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5873406Z graph_break [] 2025-12-04T09:45:06.5873478Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5873703Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5873746Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5873818Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5873856Z unimplemented [] 2025-12-04T09:45:06.5873909Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5873951Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5873987Z graph_break [] 2025-12-04T09:45:06.5874058Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5874285Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5874326Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5874397Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5874434Z unimplemented [] 2025-12-04T09:45:06.5874487Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5874556Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5874594Z graph_break [] 2025-12-04T09:45:06.5874666Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5874891Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5874933Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5875005Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5875042Z unimplemented [] 2025-12-04T09:45:06.5875097Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5875141Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5875175Z graph_break [] 2025-12-04T09:45:06.5875248Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5875475Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5875519Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5875590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5875628Z unimplemented [] 2025-12-04T09:45:06.5875681Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5875724Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5875760Z graph_break [] 2025-12-04T09:45:06.5875833Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5876063Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5876106Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5876179Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5876217Z unimplemented [] 2025-12-04T09:45:06.5876297Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5876339Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5876374Z graph_break [] 2025-12-04T09:45:06.5876447Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5876674Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5876716Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5876827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5876864Z unimplemented [] 2025-12-04T09:45:06.5876918Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5876961Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5876996Z graph_break [] 2025-12-04T09:45:06.5877069Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5877299Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5877341Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5877412Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5877449Z unimplemented [] 2025-12-04T09:45:06.5877502Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5877545Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5877580Z graph_break [] 2025-12-04T09:45:06.5877652Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5877878Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5877948Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5878024Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5878061Z unimplemented [] 2025-12-04T09:45:06.5878116Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5878157Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5878192Z graph_break [] 2025-12-04T09:45:06.5878264Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5878491Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5878532Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5878604Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5878640Z unimplemented [] 2025-12-04T09:45:06.5878695Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5878738Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5878775Z graph_break [] 2025-12-04T09:45:06.5878846Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5879074Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5879115Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5879188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5879225Z unimplemented [] 2025-12-04T09:45:06.5879280Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5879321Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5879358Z graph_break [] 2025-12-04T09:45:06.5879430Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5879660Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5879738Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5879810Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5879846Z unimplemented [] 2025-12-04T09:45:06.5879901Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5879942Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5879978Z graph_break [] 2025-12-04T09:45:06.5880050Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5880280Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5880321Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5880393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5880431Z unimplemented [] 2025-12-04T09:45:06.5880487Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5880529Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5880565Z graph_break [] 2025-12-04T09:45:06.5880637Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5880863Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5880904Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5880976Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5881012Z unimplemented [] 2025-12-04T09:45:06.5881067Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5881109Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5881144Z graph_break [] 2025-12-04T09:45:06.5881236Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5881465Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5881506Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5881609Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5881654Z Traceback (most recent call last): 2025-12-04T09:45:06.5881773Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5881812Z return value(self) 2025-12-04T09:45:06.5881960Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5882010Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5882149Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5882209Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5882371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5882443Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5882496Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5882498Z 2025-12-04T09:45:06.5882540Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5882643Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5882745Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5882747Z 2025-12-04T09:45:06.5882820Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5883005Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5883008Z 2025-12-04T09:45:06.5883118Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5883191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5883228Z unimplemented [] 2025-12-04T09:45:06.5883285Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5883327Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5883363Z graph_break [] 2025-12-04T09:45:06.5883435Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5883664Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5883706Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5883778Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5883817Z unimplemented [] 2025-12-04T09:45:06.5883872Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5883915Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5883951Z graph_break [] 2025-12-04T09:45:06.5884023Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5884252Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5884293Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5884365Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5884402Z unimplemented [] 2025-12-04T09:45:06.5884456Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5884498Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5884533Z graph_break [] 2025-12-04T09:45:06.5884625Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5884854Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5884897Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5884969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5885006Z unimplemented [] 2025-12-04T09:45:06.5885061Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5885103Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5885139Z graph_break [] 2025-12-04T09:45:06.5885210Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5885436Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5885477Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5885551Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5885589Z unimplemented [] 2025-12-04T09:45:06.5885644Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5885686Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5885721Z graph_break [] 2025-12-04T09:45:06.5885793Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5886019Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5886060Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5886131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5886168Z unimplemented [] 2025-12-04T09:45:06.5886222Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5886264Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5886301Z graph_break [] 2025-12-04T09:45:06.5886397Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5886627Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5886668Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5886781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5886817Z unimplemented [] 2025-12-04T09:45:06.5886871Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5886914Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5886951Z graph_break [] 2025-12-04T09:45:06.5887022Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5887251Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5887296Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5887369Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5887405Z unimplemented [] 2025-12-04T09:45:06.5887460Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5887501Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5887537Z graph_break [] 2025-12-04T09:45:06.5887609Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5887835Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5887876Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5887948Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5887984Z unimplemented [] 2025-12-04T09:45:06.5888081Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5888126Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5888164Z graph_break [] 2025-12-04T09:45:06.5888237Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5888464Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5888505Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5888578Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5888614Z unimplemented [] 2025-12-04T09:45:06.5888671Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5888714Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5888749Z graph_break [] 2025-12-04T09:45:06.5888822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5889051Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5889093Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5889166Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5889202Z unimplemented [] 2025-12-04T09:45:06.5889258Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5889299Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5889335Z graph_break [] 2025-12-04T09:45:06.5889408Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5889636Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5889678Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5889752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5889819Z unimplemented [] 2025-12-04T09:45:06.5889874Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5889916Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5889952Z graph_break [] 2025-12-04T09:45:06.5890024Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5890251Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5890292Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5890365Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5890403Z unimplemented [] 2025-12-04T09:45:06.5890459Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5890503Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5890540Z graph_break [] 2025-12-04T09:45:06.5890613Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5890844Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5890884Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5890959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5890996Z unimplemented [] 2025-12-04T09:45:06.5891052Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5891094Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5891132Z graph_break [] 2025-12-04T09:45:06.5891204Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5891453Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5891498Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5891569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5891608Z unimplemented [] 2025-12-04T09:45:06.5891662Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5891705Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5891742Z graph_break [] 2025-12-04T09:45:06.5891815Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5892040Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5892082Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5892154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5892194Z unimplemented [] 2025-12-04T09:45:06.5892249Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5892299Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5892334Z graph_break [] 2025-12-04T09:45:06.5892407Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5892636Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5892677Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5892748Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5892785Z unimplemented [] 2025-12-04T09:45:06.5892839Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5892882Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5892918Z graph_break [] 2025-12-04T09:45:06.5892990Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5893217Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5893290Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5893363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5893403Z unimplemented [] 2025-12-04T09:45:06.5893457Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5893499Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5893533Z graph_break [] 2025-12-04T09:45:06.5893607Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5893832Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5893875Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5893948Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5893986Z unimplemented [] 2025-12-04T09:45:06.5894042Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5894085Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5894121Z graph_break [] 2025-12-04T09:45:06.5894196Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5894423Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5894466Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5894538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5894576Z unimplemented [] 2025-12-04T09:45:06.5894632Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5894673Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5894708Z graph_break [] 2025-12-04T09:45:06.5894801Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5895029Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5895073Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5895144Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5895181Z unimplemented [] 2025-12-04T09:45:06.5895235Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5895279Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5895315Z graph_break [] 2025-12-04T09:45:06.5895389Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5895616Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5895657Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5895732Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5895771Z unimplemented [] 2025-12-04T09:45:06.5895825Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5895867Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5895902Z graph_break [] 2025-12-04T09:45:06.5895973Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5896200Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5896244Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5896316Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5896354Z unimplemented [] 2025-12-04T09:45:06.5896409Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5896477Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5896514Z graph_break [] 2025-12-04T09:45:06.5896587Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5896862Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5896905Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5896978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5897017Z unimplemented [] 2025-12-04T09:45:06.5897071Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5897114Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5897150Z graph_break [] 2025-12-04T09:45:06.5897225Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5897454Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5897500Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5897571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5897610Z unimplemented [] 2025-12-04T09:45:06.5897664Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5897707Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5897741Z graph_break [] 2025-12-04T09:45:06.5897817Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5898042Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5898085Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5898186Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5898228Z unimplemented [] 2025-12-04T09:45:06.5898284Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5898327Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5898362Z graph_break [] 2025-12-04T09:45:06.5898434Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5898660Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5898702Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5898775Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5898813Z unimplemented [] 2025-12-04T09:45:06.5898867Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5898910Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5898945Z graph_break [] 2025-12-04T09:45:06.5899020Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5899250Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5899294Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5899365Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5899403Z unimplemented [] 2025-12-04T09:45:06.5899458Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5899502Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5899538Z graph_break [] 2025-12-04T09:45:06.5899611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5899840Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5899883Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5899988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5900025Z unimplemented [] 2025-12-04T09:45:06.5900081Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5900124Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5900158Z graph_break [] 2025-12-04T09:45:06.5900233Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5900460Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5900502Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5900575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5900612Z unimplemented [] 2025-12-04T09:45:06.5900671Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5900713Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5900753Z graph_break [] 2025-12-04T09:45:06.5900826Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5901057Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5901098Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5901173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5901211Z unimplemented [] 2025-12-04T09:45:06.5901265Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5901307Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5901345Z graph_break [] 2025-12-04T09:45:06.5901417Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5901662Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5901707Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5901783Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5901820Z unimplemented [] 2025-12-04T09:45:06.5901879Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5901921Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5901958Z graph_break [] 2025-12-04T09:45:06.5902031Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5902259Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5902299Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5902375Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5902411Z unimplemented [] 2025-12-04T09:45:06.5902467Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5902509Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5902545Z graph_break [] 2025-12-04T09:45:06.5902617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5902842Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5902884Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5902959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5902997Z unimplemented [] 2025-12-04T09:45:06.5903055Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5903096Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5903135Z graph_break [] 2025-12-04T09:45:06.5903207Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5903457Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5903497Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5903570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5903608Z unimplemented [] 2025-12-04T09:45:06.5903665Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5903707Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5903744Z graph_break [] 2025-12-04T09:45:06.5903817Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5904045Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5904088Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5904162Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5904198Z unimplemented [] 2025-12-04T09:45:06.5904252Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5904293Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5904332Z graph_break [] 2025-12-04T09:45:06.5904405Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5904630Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5904672Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5904746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5904784Z unimplemented [] 2025-12-04T09:45:06.5904841Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5904908Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5904949Z graph_break [] 2025-12-04T09:45:06.5905022Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5905255Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5905297Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5905370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5905408Z unimplemented [] 2025-12-04T09:45:06.5905465Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5905507Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5905544Z graph_break [] 2025-12-04T09:45:06.5905615Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5905845Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5905890Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5905962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5905998Z unimplemented [] 2025-12-04T09:45:06.5906055Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5906098Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5906136Z graph_break [] 2025-12-04T09:45:06.5906208Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5906434Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5906475Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5906549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5906585Z unimplemented [] 2025-12-04T09:45:06.5906664Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5906706Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5906780Z graph_break [] 2025-12-04T09:45:06.5906853Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5907081Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5907123Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5907197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5907234Z unimplemented [] 2025-12-04T09:45:06.5907290Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5907331Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5907370Z graph_break [] 2025-12-04T09:45:06.5907443Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5907672Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5907713Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5907788Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5907826Z unimplemented [] 2025-12-04T09:45:06.5907882Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5907925Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5907961Z graph_break [] 2025-12-04T09:45:06.5908033Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5908260Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5908331Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5908438Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5908489Z Traceback (most recent call last): 2025-12-04T09:45:06.5908607Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5908647Z return value(self) 2025-12-04T09:45:06.5908795Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5908848Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5908985Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5909046Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5909204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5909280Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5909334Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5909336Z 2025-12-04T09:45:06.5909379Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5909481Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5909583Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5909585Z 2025-12-04T09:45:06.5909657Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5909841Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5909843Z 2025-12-04T09:45:06.5909929Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5910004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5910042Z unimplemented [] 2025-12-04T09:45:06.5910098Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5910178Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5910215Z graph_break [] 2025-12-04T09:45:06.5910287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5910515Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5910557Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5910631Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5910667Z unimplemented [] 2025-12-04T09:45:06.5910725Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5910767Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5910804Z graph_break [] 2025-12-04T09:45:06.5910878Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5911111Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5911154Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5911226Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5911264Z unimplemented [] 2025-12-04T09:45:06.5911319Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5911361Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5911397Z graph_break [] 2025-12-04T09:45:06.5911470Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5911697Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5911759Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5911833Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5911871Z unimplemented [] 2025-12-04T09:45:06.5911926Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5911967Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5912003Z graph_break [] 2025-12-04T09:45:06.5912075Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5912301Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5912343Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5912415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5912452Z unimplemented [] 2025-12-04T09:45:06.5912508Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5912551Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5912589Z graph_break [] 2025-12-04T09:45:06.5912662Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5912889Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5912932Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5913003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5913041Z unimplemented [] 2025-12-04T09:45:06.5913095Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5913137Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5913172Z graph_break [] 2025-12-04T09:45:06.5913243Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5913474Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5913539Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5913612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5913650Z unimplemented [] 2025-12-04T09:45:06.5913704Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5913746Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5913782Z graph_break [] 2025-12-04T09:45:06.5913855Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5914081Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5914123Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5914195Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5914236Z unimplemented [] 2025-12-04T09:45:06.5914290Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5914334Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5914370Z graph_break [] 2025-12-04T09:45:06.5914443Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5914668Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5914710Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5914782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5914820Z unimplemented [] 2025-12-04T09:45:06.5914874Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5914916Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5914951Z graph_break [] 2025-12-04T09:45:06.5915046Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5915272Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5915316Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5915387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5915424Z unimplemented [] 2025-12-04T09:45:06.5915479Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5915523Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5915558Z graph_break [] 2025-12-04T09:45:06.5915630Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5915856Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5915898Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5915971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5916011Z unimplemented [] 2025-12-04T09:45:06.5916065Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5916108Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5916144Z graph_break [] 2025-12-04T09:45:06.5916216Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5916441Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5916483Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5916553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5916590Z unimplemented [] 2025-12-04T09:45:06.5916644Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5916687Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5916724Z graph_break [] 2025-12-04T09:45:06.5916851Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5917077Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5917118Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5917190Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5917227Z unimplemented [] 2025-12-04T09:45:06.5917282Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5917325Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5917361Z graph_break [] 2025-12-04T09:45:06.5917433Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5917661Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5917705Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5917776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5917813Z unimplemented [] 2025-12-04T09:45:06.5917867Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5917910Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5917945Z graph_break [] 2025-12-04T09:45:06.5918017Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5918243Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5918285Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5918357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5918394Z unimplemented [] 2025-12-04T09:45:06.5918486Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5918532Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5918566Z graph_break [] 2025-12-04T09:45:06.5918639Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5918864Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5918906Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5918978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5919015Z unimplemented [] 2025-12-04T09:45:06.5919069Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5919112Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5919148Z graph_break [] 2025-12-04T09:45:06.5919220Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5919447Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5919493Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5919564Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5919601Z unimplemented [] 2025-12-04T09:45:06.5919656Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5919699Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5919734Z graph_break [] 2025-12-04T09:45:06.5919807Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5920034Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5920077Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5920153Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5920215Z unimplemented [] 2025-12-04T09:45:06.5920270Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5920313Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5920348Z graph_break [] 2025-12-04T09:45:06.5920421Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5920645Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5920688Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5920759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5920798Z unimplemented [] 2025-12-04T09:45:06.5920852Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5920895Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5920932Z graph_break [] 2025-12-04T09:45:06.5921005Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5921231Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5921274Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5921346Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5921382Z unimplemented [] 2025-12-04T09:45:06.5921435Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5921479Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5921515Z graph_break [] 2025-12-04T09:45:06.5921587Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5921838Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5921882Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5921954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5921991Z unimplemented [] 2025-12-04T09:45:06.5922046Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5922090Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5922126Z graph_break [] 2025-12-04T09:45:06.5922198Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5922424Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5922464Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5922536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5922573Z unimplemented [] 2025-12-04T09:45:06.5922629Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5922672Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5922708Z graph_break [] 2025-12-04T09:45:06.5922779Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5923003Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5923044Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5923116Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5923152Z unimplemented [] 2025-12-04T09:45:06.5923207Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5923249Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5923285Z graph_break [] 2025-12-04T09:45:06.5923356Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5923584Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5923647Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5923720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5923757Z unimplemented [] 2025-12-04T09:45:06.5923812Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5923854Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5923890Z graph_break [] 2025-12-04T09:45:06.5923962Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5924189Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5924231Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5924304Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5924342Z unimplemented [] 2025-12-04T09:45:06.5924396Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5924438Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5924474Z graph_break [] 2025-12-04T09:45:06.5924544Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5924771Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5924812Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5924884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5924920Z unimplemented [] 2025-12-04T09:45:06.5924975Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5925017Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5925072Z graph_break [] 2025-12-04T09:45:06.5925144Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5925373Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5925413Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5925486Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5925522Z unimplemented [] 2025-12-04T09:45:06.5925577Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5925619Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5925655Z graph_break [] 2025-12-04T09:45:06.5925726Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5925959Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5926002Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5926076Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5926112Z unimplemented [] 2025-12-04T09:45:06.5926167Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5926210Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5926246Z graph_break [] 2025-12-04T09:45:06.5926317Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5927808Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5927852Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5927927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5927964Z unimplemented [] 2025-12-04T09:45:06.5928025Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5928107Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5928144Z graph_break [] 2025-12-04T09:45:06.5928216Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5928446Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5928487Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5928562Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5928599Z unimplemented [] 2025-12-04T09:45:06.5928655Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5928697Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5928734Z graph_break [] 2025-12-04T09:45:06.5928804Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5929034Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5929080Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5929154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5929191Z unimplemented [] 2025-12-04T09:45:06.5929247Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5929289Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5929326Z graph_break [] 2025-12-04T09:45:06.5929399Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5929719Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5929760Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5929865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5929908Z unimplemented [] 2025-12-04T09:45:06.5929963Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5930005Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5930041Z graph_break [] 2025-12-04T09:45:06.5930114Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5930342Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5930383Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5930455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5930492Z unimplemented [] 2025-12-04T09:45:06.5930546Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5930588Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5930624Z graph_break [] 2025-12-04T09:45:06.5930700Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5930930Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5930971Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5931044Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5931081Z unimplemented [] 2025-12-04T09:45:06.5931136Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5931179Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5931215Z graph_break [] 2025-12-04T09:45:06.5931287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5931522Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5931563Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5931670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5931707Z unimplemented [] 2025-12-04T09:45:06.5931762Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5931804Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5931843Z graph_break [] 2025-12-04T09:45:06.5931914Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5932140Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5932182Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5932253Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5932291Z unimplemented [] 2025-12-04T09:45:06.5932348Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5932391Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5932427Z graph_break [] 2025-12-04T09:45:06.5932499Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5932725Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5932768Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5932839Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5932878Z unimplemented [] 2025-12-04T09:45:06.5932932Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5932976Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5933011Z graph_break [] 2025-12-04T09:45:06.5933084Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5933333Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5933379Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5933451Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5933489Z unimplemented [] 2025-12-04T09:45:06.5933543Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5933586Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5933621Z graph_break [] 2025-12-04T09:45:06.5933693Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5933917Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5933960Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5934033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5934072Z unimplemented [] 2025-12-04T09:45:06.5934127Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5934170Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5934205Z graph_break [] 2025-12-04T09:45:06.5934277Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5934502Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5934544Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5934615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5934652Z unimplemented [] 2025-12-04T09:45:06.5934707Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5934750Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5934786Z graph_break [] 2025-12-04T09:45:06.5934861Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5935110Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5935153Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5935224Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5935261Z unimplemented [] 2025-12-04T09:45:06.5935316Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5935359Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5935395Z graph_break [] 2025-12-04T09:45:06.5935468Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5935695Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5935739Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5935812Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5935849Z unimplemented [] 2025-12-04T09:45:06.5935904Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5935947Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5935983Z graph_break [] 2025-12-04T09:45:06.5936056Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5936282Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5936325Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5936396Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5936433Z unimplemented [] 2025-12-04T09:45:06.5936512Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5936556Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5936593Z graph_break [] 2025-12-04T09:45:06.5936667Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5936929Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5936972Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5937075Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5937121Z Traceback (most recent call last): 2025-12-04T09:45:06.5937242Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5937281Z return value(self) 2025-12-04T09:45:06.5937431Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5937488Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5937628Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5937689Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5937849Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5937924Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5937976Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5937979Z 2025-12-04T09:45:06.5938023Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5938126Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5938230Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5938231Z 2025-12-04T09:45:06.5938308Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5938493Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5938532Z 2025-12-04T09:45:06.5938620Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5938694Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5938733Z unimplemented [] 2025-12-04T09:45:06.5938787Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5938831Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5938867Z graph_break [] 2025-12-04T09:45:06.5938940Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5939170Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5939216Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5939288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5939328Z unimplemented [] 2025-12-04T09:45:06.5939383Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5939426Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5939462Z graph_break [] 2025-12-04T09:45:06.5939535Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5939766Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5939809Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5939881Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5939918Z unimplemented [] 2025-12-04T09:45:06.5939973Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5940040Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5940076Z graph_break [] 2025-12-04T09:45:06.5940151Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5940375Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5940418Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5940489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5940528Z unimplemented [] 2025-12-04T09:45:06.5940583Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5940626Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5940662Z graph_break [] 2025-12-04T09:45:06.5940737Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5940964Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5941008Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5941080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5941122Z unimplemented [] 2025-12-04T09:45:06.5941177Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5941221Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5941258Z graph_break [] 2025-12-04T09:45:06.5941331Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5941557Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5941602Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5941675Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5941716Z unimplemented [] 2025-12-04T09:45:06.5941774Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5941847Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5941883Z graph_break [] 2025-12-04T09:45:06.5941958Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5942185Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5942230Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5942301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5942341Z unimplemented [] 2025-12-04T09:45:06.5942395Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5942439Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5942475Z graph_break [] 2025-12-04T09:45:06.5942550Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5942776Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5942821Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5942893Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5942931Z unimplemented [] 2025-12-04T09:45:06.5942987Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5943032Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5943067Z graph_break [] 2025-12-04T09:45:06.5943143Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5943368Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5943437Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5943511Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5943552Z unimplemented [] 2025-12-04T09:45:06.5943606Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5943650Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5943686Z graph_break [] 2025-12-04T09:45:06.5943760Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5943990Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5944035Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5944107Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5944145Z unimplemented [] 2025-12-04T09:45:06.5944203Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5944251Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5944287Z graph_break [] 2025-12-04T09:45:06.5944364Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5944590Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5944633Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5944705Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5944743Z unimplemented [] 2025-12-04T09:45:06.5944797Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5944842Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5944878Z graph_break [] 2025-12-04T09:45:06.5944951Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5945183Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5945264Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5945337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5945374Z unimplemented [] 2025-12-04T09:45:06.5945430Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5945472Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5945507Z graph_break [] 2025-12-04T09:45:06.5945583Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5945812Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5945852Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5945927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5945966Z unimplemented [] 2025-12-04T09:45:06.5946022Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5946067Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5946105Z graph_break [] 2025-12-04T09:45:06.5946177Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5946404Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5946447Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5946520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5946556Z unimplemented [] 2025-12-04T09:45:06.5946614Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5946656Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5946694Z graph_break [] 2025-12-04T09:45:06.5946835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5947064Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5947108Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5947184Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5947220Z unimplemented [] 2025-12-04T09:45:06.5947280Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5947325Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5947365Z graph_break [] 2025-12-04T09:45:06.5947438Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5947669Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5947711Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5947787Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5947825Z unimplemented [] 2025-12-04T09:45:06.5947887Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5947931Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5947976Z graph_break [] 2025-12-04T09:45:06.5948052Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5948284Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5948326Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5948400Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5948440Z unimplemented [] 2025-12-04T09:45:06.5948498Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5948541Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5948582Z graph_break [] 2025-12-04T09:45:06.5948690Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5948918Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5948959Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5949038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5949079Z unimplemented [] 2025-12-04T09:45:06.5949139Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5949185Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5949224Z graph_break [] 2025-12-04T09:45:06.5949296Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5949524Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5949572Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5949647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5949686Z unimplemented [] 2025-12-04T09:45:06.5949747Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5949790Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5949828Z graph_break [] 2025-12-04T09:45:06.5949900Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5950130Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5950175Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5950249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5950290Z unimplemented [] 2025-12-04T09:45:06.5950368Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5950416Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5950451Z graph_break [] 2025-12-04T09:45:06.5950527Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5950756Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5950802Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5950873Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5950913Z unimplemented [] 2025-12-04T09:45:06.5950968Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5951013Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5951050Z graph_break [] 2025-12-04T09:45:06.5951125Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5951354Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5951399Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5951471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5951509Z unimplemented [] 2025-12-04T09:45:06.5951563Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5951606Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5951641Z graph_break [] 2025-12-04T09:45:06.5951716Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5951944Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5951989Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5952066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5952136Z unimplemented [] 2025-12-04T09:45:06.5952191Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5952237Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5952273Z graph_break [] 2025-12-04T09:45:06.5952350Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5952581Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5952628Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5952704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5952745Z unimplemented [] 2025-12-04T09:45:06.5952801Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5952847Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5952888Z graph_break [] 2025-12-04T09:45:06.5952965Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5953194Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5953239Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5953315Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5953357Z unimplemented [] 2025-12-04T09:45:06.5953412Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5953460Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5953497Z graph_break [] 2025-12-04T09:45:06.5953573Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5953823Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5953879Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5953950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5953991Z unimplemented [] 2025-12-04T09:45:06.5954046Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5954091Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5954128Z graph_break [] 2025-12-04T09:45:06.5954200Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5954430Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5954473Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5954546Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5954585Z unimplemented [] 2025-12-04T09:45:06.5954641Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5954692Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5954728Z graph_break [] 2025-12-04T09:45:06.5954804Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5955030Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5955075Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5955146Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5955185Z unimplemented [] 2025-12-04T09:45:06.5955240Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5955285Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5955321Z graph_break [] 2025-12-04T09:45:06.5955392Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5955619Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5955691Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5955763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5955803Z unimplemented [] 2025-12-04T09:45:06.5955858Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5955903Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5955940Z graph_break [] 2025-12-04T09:45:06.5956015Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5956240Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5956286Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5956360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5956402Z unimplemented [] 2025-12-04T09:45:06.5956457Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5956503Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5956539Z graph_break [] 2025-12-04T09:45:06.5956614Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5956871Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5956915Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5956987Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5957027Z unimplemented [] 2025-12-04T09:45:06.5957083Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5957127Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5957200Z graph_break [] 2025-12-04T09:45:06.5957278Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5957505Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5957551Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5957624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5957664Z unimplemented [] 2025-12-04T09:45:06.5957719Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5957765Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5957802Z graph_break [] 2025-12-04T09:45:06.5957877Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5958105Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5958153Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5958225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5958266Z unimplemented [] 2025-12-04T09:45:06.5958322Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5958369Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5958405Z graph_break [] 2025-12-04T09:45:06.5958481Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5958711Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5958759Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5958832Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5958871Z unimplemented [] 2025-12-04T09:45:06.5958929Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5959005Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5959040Z graph_break [] 2025-12-04T09:45:06.5959114Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5959339Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5959383Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5959456Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5959496Z unimplemented [] 2025-12-04T09:45:06.5959550Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5959593Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5959628Z graph_break [] 2025-12-04T09:45:06.5959700Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5959927Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5959972Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5960043Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5960082Z unimplemented [] 2025-12-04T09:45:06.5960137Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5960184Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5960220Z graph_break [] 2025-12-04T09:45:06.5960295Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5960523Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5960568Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5960665Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5960707Z unimplemented [] 2025-12-04T09:45:06.5960764Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5960810Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5960846Z graph_break [] 2025-12-04T09:45:06.5960921Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5961149Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5961194Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5961267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5961306Z unimplemented [] 2025-12-04T09:45:06.5961361Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5961407Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5961444Z graph_break [] 2025-12-04T09:45:06.5961522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5961749Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5961794Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5961866Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5961904Z unimplemented [] 2025-12-04T09:45:06.5961959Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5962003Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5962038Z graph_break [] 2025-12-04T09:45:06.5962112Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5962340Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5962383Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5962480Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5962517Z unimplemented [] 2025-12-04T09:45:06.5962573Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5962615Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5962650Z graph_break [] 2025-12-04T09:45:06.5962723Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5962952Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5962994Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5963066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5963103Z unimplemented [] 2025-12-04T09:45:06.5963161Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5963203Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5963241Z graph_break [] 2025-12-04T09:45:06.5963313Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5963542Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5963583Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5963657Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5963694Z unimplemented [] 2025-12-04T09:45:06.5963749Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5963792Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5963829Z graph_break [] 2025-12-04T09:45:06.5963900Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5964149Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5964193Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5964266Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5964302Z unimplemented [] 2025-12-04T09:45:06.5964357Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5964399Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5964436Z graph_break [] 2025-12-04T09:45:06.5964507Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5964734Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5964775Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5964851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5964890Z unimplemented [] 2025-12-04T09:45:06.5964946Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5964988Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5965025Z graph_break [] 2025-12-04T09:45:06.5965098Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5965325Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5965366Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5965439Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5965475Z unimplemented [] 2025-12-04T09:45:06.5965532Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5965574Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5965610Z graph_break [] 2025-12-04T09:45:06.5965683Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5965941Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5965982Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5966054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5966090Z unimplemented [] 2025-12-04T09:45:06.5966146Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5966189Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5966225Z graph_break [] 2025-12-04T09:45:06.5966297Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5966525Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5966568Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5966641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5966678Z unimplemented [] 2025-12-04T09:45:06.5966734Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5966902Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5966939Z graph_break [] 2025-12-04T09:45:06.5967011Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5967240Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5967281Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5967353Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5967389Z unimplemented [] 2025-12-04T09:45:06.5967479Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5967522Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5967561Z graph_break [] 2025-12-04T09:45:06.5967633Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5967859Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5967900Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5967972Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5968009Z unimplemented [] 2025-12-04T09:45:06.5968065Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5968107Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5968143Z graph_break [] 2025-12-04T09:45:06.5968214Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5968442Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5968484Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5968589Z ____ AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda _____ 2025-12-04T09:45:06.5968635Z Traceback (most recent call last): 2025-12-04T09:45:06.5968755Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:45:06.5968794Z return value(self) 2025-12-04T09:45:06.5968944Z File "/var/lib/jenkins/pytorch/test/inductor/test_aot_inductor.py", line 7544, in test_copy_non_blocking_is_pinned 2025-12-04T09:45:06.5968995Z self.assertEqual(outputs, outputs_aoti) 2025-12-04T09:45:06.5969133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:45:06.5969192Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:45:06.5969357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:45:06.5969459Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:45:06.5969512Z AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.5969514Z 2025-12-04T09:45:06.5969558Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.5969659Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.5969761Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.5969763Z 2025-12-04T09:45:06.5969835Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.5970018Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.5970020Z 2025-12-04T09:45:06.5970109Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.5970182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5970221Z unimplemented [] 2025-12-04T09:45:06.5970278Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5970320Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5970357Z graph_break [] 2025-12-04T09:45:06.5970429Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5970657Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5970699Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5970772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5970809Z unimplemented [] 2025-12-04T09:45:06.5970865Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5970935Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5970972Z graph_break [] 2025-12-04T09:45:06.5971045Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5971275Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5971316Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5971390Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5971427Z unimplemented [] 2025-12-04T09:45:06.5971483Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5971525Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5971561Z graph_break [] 2025-12-04T09:45:06.5971633Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5971861Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5971905Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5971977Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5972014Z unimplemented [] 2025-12-04T09:45:06.5972070Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5972112Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5972149Z graph_break [] 2025-12-04T09:45:06.5972220Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5972447Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5972488Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5972561Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5972600Z unimplemented [] 2025-12-04T09:45:06.5972657Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5972721Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5972757Z graph_break [] 2025-12-04T09:45:06.5972829Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5973054Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5973095Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5973168Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5973204Z unimplemented [] 2025-12-04T09:45:06.5973260Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5973301Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5973338Z graph_break [] 2025-12-04T09:45:06.5973413Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5973642Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5973685Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5973759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5973796Z unimplemented [] 2025-12-04T09:45:06.5973851Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5973893Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5973930Z graph_break [] 2025-12-04T09:45:06.5974002Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5974228Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5974291Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5974365Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5974404Z unimplemented [] 2025-12-04T09:45:06.5974460Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5974502Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5974538Z graph_break [] 2025-12-04T09:45:06.5974609Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5974837Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5974878Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5974951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5974989Z unimplemented [] 2025-12-04T09:45:06.5975045Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5975089Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5975127Z graph_break [] 2025-12-04T09:45:06.5975200Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5975426Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5975469Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5975541Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5975579Z unimplemented [] 2025-12-04T09:45:06.5975633Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5975674Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5975711Z graph_break [] 2025-12-04T09:45:06.5975783Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5976012Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5976078Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5976150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5976188Z unimplemented [] 2025-12-04T09:45:06.5976242Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5976285Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5976321Z graph_break [] 2025-12-04T09:45:06.5976394Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5976617Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5976660Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5976731Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5976800Z unimplemented [] 2025-12-04T09:45:06.5976856Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5976901Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5976936Z graph_break [] 2025-12-04T09:45:06.5977010Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5977236Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5977278Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5977350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5977387Z unimplemented [] 2025-12-04T09:45:06.5977441Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5977485Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5977520Z graph_break [] 2025-12-04T09:45:06.5977626Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5977854Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5977898Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5977969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5978007Z unimplemented [] 2025-12-04T09:45:06.5978061Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5978104Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5978139Z graph_break [] 2025-12-04T09:45:06.5978211Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5978436Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5978478Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5978551Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5978589Z unimplemented [] 2025-12-04T09:45:06.5978644Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5978687Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5978722Z graph_break [] 2025-12-04T09:45:06.5978794Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5979020Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5979061Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5979132Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5979169Z unimplemented [] 2025-12-04T09:45:06.5979224Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5979267Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5979306Z graph_break [] 2025-12-04T09:45:06.5979411Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5979637Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5979680Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5979754Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5979794Z unimplemented [] 2025-12-04T09:45:06.5979848Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5979892Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5979927Z graph_break [] 2025-12-04T09:45:06.5980001Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5980230Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5980274Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5980345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5980383Z unimplemented [] 2025-12-04T09:45:06.5980437Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5980481Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5980516Z graph_break [] 2025-12-04T09:45:06.5980589Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5980816Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5980859Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5980932Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5980970Z unimplemented [] 2025-12-04T09:45:06.5981045Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5981091Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5981127Z graph_break [] 2025-12-04T09:45:06.5981200Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5981426Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5981470Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5981542Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5981580Z unimplemented [] 2025-12-04T09:45:06.5981635Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5981680Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5981715Z graph_break [] 2025-12-04T09:45:06.5981790Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5982019Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5982064Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5982135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5982173Z unimplemented [] 2025-12-04T09:45:06.5982227Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5982271Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5982305Z graph_break [] 2025-12-04T09:45:06.5982379Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5982603Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5982646Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5982719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5982787Z unimplemented [] 2025-12-04T09:45:06.5982841Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5982883Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5982918Z graph_break [] 2025-12-04T09:45:06.5982992Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5983217Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5983259Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5983331Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5983368Z unimplemented [] 2025-12-04T09:45:06.5983423Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5983466Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5983503Z graph_break [] 2025-12-04T09:45:06.5983576Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5983805Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5983848Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5983920Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5983958Z unimplemented [] 2025-12-04T09:45:06.5984012Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5984057Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5984092Z graph_break [] 2025-12-04T09:45:06.5984165Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5984418Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5984462Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5984535Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5984572Z unimplemented [] 2025-12-04T09:45:06.5984628Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5984671Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5984708Z graph_break [] 2025-12-04T09:45:06.5984779Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5985005Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5985046Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5985119Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5985156Z unimplemented [] 2025-12-04T09:45:06.5985213Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5985257Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5985293Z graph_break [] 2025-12-04T09:45:06.5985365Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5985592Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5985633Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5985706Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5985742Z unimplemented [] 2025-12-04T09:45:06.5985798Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5985840Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5985877Z graph_break [] 2025-12-04T09:45:06.5985949Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5986179Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5986249Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5986323Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5986360Z unimplemented [] 2025-12-04T09:45:06.5986415Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5986457Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5986494Z graph_break [] 2025-12-04T09:45:06.5986565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5986822Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5986864Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5986939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5986979Z unimplemented [] 2025-12-04T09:45:06.5987035Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5987077Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5987113Z graph_break [] 2025-12-04T09:45:06.5987186Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5987414Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5987455Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5987529Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5987566Z unimplemented [] 2025-12-04T09:45:06.5987622Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5987664Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5987736Z graph_break [] 2025-12-04T09:45:06.5987809Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5988042Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5988083Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5988157Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5988193Z unimplemented [] 2025-12-04T09:45:06.5988249Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5988292Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5988328Z graph_break [] 2025-12-04T09:45:06.5988400Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5988629Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5988672Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5988746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5988782Z unimplemented [] 2025-12-04T09:45:06.5988837Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5988880Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5988916Z graph_break [] 2025-12-04T09:45:06.5988988Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5989213Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5989255Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5989328Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5989365Z unimplemented [] 2025-12-04T09:45:06.5989423Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5989494Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5989531Z graph_break [] 2025-12-04T09:45:06.5989602Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5989828Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5989869Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5989943Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5989980Z unimplemented [] 2025-12-04T09:45:06.5990036Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5990078Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5990115Z graph_break [] 2025-12-04T09:45:06.5990186Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5990415Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5990458Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5990531Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5990568Z unimplemented [] 2025-12-04T09:45:06.5990624Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5990665Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5990703Z graph_break [] 2025-12-04T09:45:06.5990775Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5991002Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5991044Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5991141Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5991180Z unimplemented [] 2025-12-04T09:45:06.5991236Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5991278Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5991315Z graph_break [] 2025-12-04T09:45:06.5991387Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5991615Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5991656Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5991728Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5991765Z unimplemented [] 2025-12-04T09:45:06.5991821Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5991863Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5991900Z graph_break [] 2025-12-04T09:45:06.5991974Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5992208Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5992250Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5992324Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5992361Z unimplemented [] 2025-12-04T09:45:06.5992417Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5992460Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5992502Z graph_break [] 2025-12-04T09:45:06.5992575Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5992807Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5992851Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5992951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5992994Z unimplemented [] 2025-12-04T09:45:06.5993051Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5993097Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5993134Z graph_break [] 2025-12-04T09:45:06.5993207Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5993437Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5993481Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5993555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5993593Z unimplemented [] 2025-12-04T09:45:06.5993649Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5993696Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5993734Z graph_break [] 2025-12-04T09:45:06.5993811Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5994040Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5994085Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5994162Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5994202Z unimplemented [] 2025-12-04T09:45:06.5994257Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5994303Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5994340Z graph_break [] 2025-12-04T09:45:06.5994417Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5994668Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5994717Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5994790Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5994831Z unimplemented [] 2025-12-04T09:45:06.5994886Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5994932Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5994967Z graph_break [] 2025-12-04T09:45:06.5995044Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5995271Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5995316Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5995390Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5995434Z unimplemented [] 2025-12-04T09:45:06.5995489Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5995537Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5995573Z graph_break [] 2025-12-04T09:45:06.5995648Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5995873Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5995916Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5995989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5996029Z unimplemented [] 2025-12-04T09:45:06.5996086Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5996135Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5996170Z graph_break [] 2025-12-04T09:45:06.5996249Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5996513Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5996558Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5996630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5996672Z unimplemented [] 2025-12-04T09:45:06.5996727Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5996804Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5996842Z graph_break [] 2025-12-04T09:45:06.5996917Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5997149Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5997195Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5997269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5997309Z unimplemented [] 2025-12-04T09:45:06.5997364Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5997409Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5997446Z graph_break [] 2025-12-04T09:45:06.5997522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5997748Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5997794Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5997866Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5997907Z unimplemented [] 2025-12-04T09:45:06.5997993Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5998039Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5998077Z graph_break [] 2025-12-04T09:45:06.5998150Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5998376Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5998422Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5998496Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5998535Z unimplemented [] 2025-12-04T09:45:06.5998590Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5998638Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5998674Z graph_break [] 2025-12-04T09:45:06.5998750Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5998980Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5999028Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5999101Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5999140Z unimplemented [] 2025-12-04T09:45:06.5999196Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5999243Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5999279Z graph_break [] 2025-12-04T09:45:06.5999355Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.5999580Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.5999626Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.5999702Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:45:06.5999742Z unimplemented [] 2025-12-04T09:45:06.5999832Z stats [('calls_captured', 7), ('unique_graphs', 1)] 2025-12-04T09:45:06.5999876Z inductor [('extern_calls', 2)] 2025-12-04T09:45:06.5999912Z graph_break [] 2025-12-04T09:45:06.5999989Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:45:06.6000216Z /opt/conda/envs/py_3.10/lib/python3.10/copyreg.py:101: FutureWarning: `isinstance(treespec, LeafSpec)` is deprecated, use `isinstance(treespec, TreeSpec) and treespec.is_leaf()` instead. 2025-12-04T09:45:06.6000261Z return cls.__new__(cls, *args) 2025-12-04T09:45:06.6000484Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml - 2025-12-04T09:45:06.6000549Z =========================== short test summary info ============================ 2025-12-04T09:45:06.6000796Z FAILED [6.3382s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6000800Z 2025-12-04T09:45:06.6000844Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6000950Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6001052Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6001054Z 2025-12-04T09:45:06.6001129Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6001313Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6001315Z 2025-12-04T09:45:06.6001406Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6001642Z FAILED [3.0370s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6001645Z 2025-12-04T09:45:06.6001695Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6001797Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6001900Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6001902Z 2025-12-04T09:45:06.6001977Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6002162Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6002164Z 2025-12-04T09:45:06.6002253Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6002468Z FAILED [3.0923s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6002470Z 2025-12-04T09:45:06.6002518Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6002621Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6002724Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6002726Z 2025-12-04T09:45:06.6002799Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6002987Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6002989Z 2025-12-04T09:45:06.6003075Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6003290Z FAILED [5.8673s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6003292Z 2025-12-04T09:45:06.6003335Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6003440Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6003561Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6003563Z 2025-12-04T09:45:06.6003637Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6003818Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6003823Z 2025-12-04T09:45:06.6003907Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6004120Z FAILED [3.0095s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6004122Z 2025-12-04T09:45:06.6004166Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6004269Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6004369Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6004373Z 2025-12-04T09:45:06.6004449Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6004629Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6004631Z 2025-12-04T09:45:06.6004719Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6004930Z FAILED [3.0399s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6004932Z 2025-12-04T09:45:06.6004978Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6005078Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6005201Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6005204Z 2025-12-04T09:45:06.6005279Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6005464Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6005466Z 2025-12-04T09:45:06.6005553Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6005766Z FAILED [3.0553s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6005768Z 2025-12-04T09:45:06.6005814Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6005914Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6006015Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6006018Z 2025-12-04T09:45:06.6006090Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6006276Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6006278Z 2025-12-04T09:45:06.6006362Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6006581Z FAILED [3.0247s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6006583Z 2025-12-04T09:45:06.6006627Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6006728Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6006860Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6006864Z 2025-12-04T09:45:06.6006939Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6007123Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6007159Z 2025-12-04T09:45:06.6007245Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6007460Z FAILED [3.0609s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6007462Z 2025-12-04T09:45:06.6007506Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6007608Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6007710Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6007712Z 2025-12-04T09:45:06.6007787Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6007972Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6007976Z 2025-12-04T09:45:06.6008065Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6008278Z FAILED [8.7884s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6008279Z 2025-12-04T09:45:06.6008326Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6008426Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6008525Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6008527Z 2025-12-04T09:45:06.6008600Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6008812Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6008816Z 2025-12-04T09:45:06.6008902Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6009113Z FAILED [8.1315s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6009115Z 2025-12-04T09:45:06.6009161Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6009258Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6009357Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6009359Z 2025-12-04T09:45:06.6009430Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6009611Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6009613Z 2025-12-04T09:45:06.6009701Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6009919Z FAILED [8.0859s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6009921Z 2025-12-04T09:45:06.6009965Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6010063Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6010163Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6010165Z 2025-12-04T09:45:06.6010236Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6010419Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6010421Z 2025-12-04T09:45:06.6010506Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6010720Z FAILED [4.5393s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6010748Z 2025-12-04T09:45:06.6010791Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6010890Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6010988Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6010990Z 2025-12-04T09:45:06.6011064Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6011243Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6011245Z 2025-12-04T09:45:06.6011330Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6011543Z FAILED [6.5972s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6011548Z 2025-12-04T09:45:06.6011591Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6011690Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6011788Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6011790Z 2025-12-04T09:45:06.6011863Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6012044Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6012046Z 2025-12-04T09:45:06.6012131Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6012367Z FAILED [8.7425s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6012370Z 2025-12-04T09:45:06.6012415Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6012513Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6012611Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6012613Z 2025-12-04T09:45:06.6012683Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6012864Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6012866Z 2025-12-04T09:45:06.6012949Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6013160Z FAILED [6.4504s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6013164Z 2025-12-04T09:45:06.6013208Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6013308Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6013407Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6013409Z 2025-12-04T09:45:06.6013479Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6013660Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6013662Z 2025-12-04T09:45:06.6013745Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6013957Z FAILED [3.1455s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6013958Z 2025-12-04T09:45:06.6014002Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6014102Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6014225Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6014227Z 2025-12-04T09:45:06.6014301Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6014483Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6014487Z 2025-12-04T09:45:06.6014571Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6014785Z FAILED [3.0648s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6014787Z 2025-12-04T09:45:06.6014829Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6014931Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6015030Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6015032Z 2025-12-04T09:45:06.6015104Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6015285Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6015287Z 2025-12-04T09:45:06.6015372Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6015583Z FAILED [3.0759s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6015584Z 2025-12-04T09:45:06.6015629Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6015727Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6015850Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6015854Z 2025-12-04T09:45:06.6015926Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6016108Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6016110Z 2025-12-04T09:45:06.6016195Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6016405Z FAILED [3.0232s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6016407Z 2025-12-04T09:45:06.6016451Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6016550Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6016651Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6016653Z 2025-12-04T09:45:06.6016726Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6016949Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6016951Z 2025-12-04T09:45:06.6017034Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6017246Z FAILED [6.5313s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6017248Z 2025-12-04T09:45:06.6017291Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6017390Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6017489Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6017491Z 2025-12-04T09:45:06.6017566Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6017775Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6017779Z 2025-12-04T09:45:06.6017862Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6018074Z FAILED [3.1357s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6018075Z 2025-12-04T09:45:06.6018118Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6018219Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6018318Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6018320Z 2025-12-04T09:45:06.6018393Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6018575Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6018578Z 2025-12-04T09:45:06.6018664Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6018878Z FAILED [4.5588s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6018880Z 2025-12-04T09:45:06.6018928Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6019026Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6019126Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6019128Z 2025-12-04T09:45:06.6019199Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6019406Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6019410Z 2025-12-04T09:45:06.6019496Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6019707Z FAILED [6.6000s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6019709Z 2025-12-04T09:45:06.6019753Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6019852Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6019951Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6019953Z 2025-12-04T09:45:06.6020024Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6020206Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6020210Z 2025-12-04T09:45:06.6020294Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6020511Z FAILED [3.0815s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6020513Z 2025-12-04T09:45:06.6020555Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6020655Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6020752Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6020756Z 2025-12-04T09:45:06.6020826Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6021008Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6021010Z 2025-12-04T09:45:06.6021096Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6021337Z FAILED [4.6693s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6021339Z 2025-12-04T09:45:06.6021383Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6021483Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6021581Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6021582Z 2025-12-04T09:45:06.6021655Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6021835Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6021837Z 2025-12-04T09:45:06.6021922Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6022135Z FAILED [7.3739s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6022140Z 2025-12-04T09:45:06.6022182Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6022280Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6022379Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6022381Z 2025-12-04T09:45:06.6022454Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6022632Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6022634Z 2025-12-04T09:45:06.6022718Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6022954Z FAILED [9.1520s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6022958Z 2025-12-04T09:45:06.6023003Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6023101Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6023201Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6023202Z 2025-12-04T09:45:06.6023275Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6023458Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6023460Z 2025-12-04T09:45:06.6023544Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6023758Z FAILED [3.0411s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6023760Z 2025-12-04T09:45:06.6023806Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6023904Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6024003Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6024005Z 2025-12-04T09:45:06.6024076Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6024256Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6024258Z 2025-12-04T09:45:06.6024340Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6024552Z FAILED [4.7527s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6024554Z 2025-12-04T09:45:06.6024598Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6024724Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6024825Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6024827Z 2025-12-04T09:45:06.6024899Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6025078Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6025080Z 2025-12-04T09:45:06.6025165Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6025377Z FAILED [12.3107s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6025381Z 2025-12-04T09:45:06.6025423Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6025523Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6025622Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6025624Z 2025-12-04T09:45:06.6025696Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6025876Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6025878Z 2025-12-04T09:45:06.6025962Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6026174Z FAILED [8.0937s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6026176Z 2025-12-04T09:45:06.6026221Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6026344Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6026445Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6026448Z 2025-12-04T09:45:06.6026518Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6026705Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6026707Z 2025-12-04T09:45:06.6026829Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6027042Z FAILED [12.2574s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6027044Z 2025-12-04T09:45:06.6027088Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6027186Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6027287Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6027290Z 2025-12-04T09:45:06.6027361Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6027544Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6027546Z 2025-12-04T09:45:06.6027629Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6027841Z FAILED [4.4793s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6027843Z 2025-12-04T09:45:06.6027886Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6027985Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6028084Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6028087Z 2025-12-04T09:45:06.6028160Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6028373Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6028377Z 2025-12-04T09:45:06.6028461Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6028671Z FAILED [8.9116s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6028673Z 2025-12-04T09:45:06.6028715Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6028814Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6028911Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6028913Z 2025-12-04T09:45:06.6028986Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6029169Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6029172Z 2025-12-04T09:45:06.6029258Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6029469Z FAILED [6.4163s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6029471Z 2025-12-04T09:45:06.6029515Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6029615Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6029714Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6029715Z 2025-12-04T09:45:06.6029786Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6029993Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6029997Z 2025-12-04T09:45:06.6030083Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6030294Z FAILED [3.0219s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6030296Z 2025-12-04T09:45:06.6030341Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6030439Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6030537Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6030539Z 2025-12-04T09:45:06.6030612Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6030796Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6030799Z 2025-12-04T09:45:06.6030886Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6031097Z FAILED [8.4732s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6031099Z 2025-12-04T09:45:06.6031142Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6031241Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6031338Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6031340Z 2025-12-04T09:45:06.6031412Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6031591Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6031594Z 2025-12-04T09:45:06.6031680Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6031916Z FAILED [3.0237s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6031917Z 2025-12-04T09:45:06.6031960Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6032059Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6032156Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6032158Z 2025-12-04T09:45:06.6032229Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6032409Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6032411Z 2025-12-04T09:45:06.6032497Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6032707Z FAILED [3.0111s] inductor/test_aot_inductor.py::AOTInductorTestABICompatibleGpu::test_copy_non_blocking_is_pinned_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:45:06.6032711Z 2025-12-04T09:45:06.6032754Z Mismatched elements: 4 / 8 (50.0%) 2025-12-04T09:45:06.6032853Z Greatest absolute difference: 2.2495734691619873 at index (2, 0) (up to 1e-05 allowed) 2025-12-04T09:45:06.6032951Z Greatest relative difference: 2.118741035461426 at index (3, 1) (up to 1.3e-06 allowed) 2025-12-04T09:45:06.6032953Z 2025-12-04T09:45:06.6033026Z To execute this test, run the following from the base repo dir: 2025-12-04T09:45:06.6033205Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_aot_inductor.py AOTInductorTestABICompatibleGpu.test_copy_non_blocking_is_pinned_cuda 2025-12-04T09:45:06.6033207Z 2025-12-04T09:45:06.6033291Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:45:06.6033382Z ================== 40 failed, 110 passed in 702.51s (0:11:42) ================== 2025-12-04T09:45:06.6033386Z 2025-12-04T09:45:06.6033555Z FINISHED PRINTING LOG FILE of inductor/test_aot_inductor 3/3 (test/test-reports/inductor.test_aot_inductor_3.3_f0c7cbe1a17c7fbf_.log) 2025-12-04T09:45:06.6033558Z 2025-12-04T09:45:06.6033675Z Finished inductor/test_aot_inductor 3/3 ... [2025-12-04 09:45:06.500249][2225330.957563452], took 11.87min 2025-12-04T09:45:06.6033907Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:45:08.7423019Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:45:08.7423436Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-12-04T09:45:08.7423715Z Uploading artifacts took 0.00 seconds 2025-12-04T09:45:08.7423921Z inductor/test_aot_inductor 3/3 failed! 2025-12-04T09:45:08.7424403Z Running inductor/test_torchinductor_codegen_dynamic_shapes 2/4 ... [2025-12-04 09:45:08.742297][2225333.199615614] 2025-12-04T09:45:08.7424783Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:45:08.7427299Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_codegen_dynamic_shapes.py', '--shard-id=2', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:45:08.742583] 2025-12-04T09:45:19.5492969Z 2025-12-04T09:45:19.5494630Z inductor/test_torchinductor_codegen_dynamic_shapes 2/4 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_codegen_dynamic_shapes_2.4_e4cf46a75bdd8b6a_.log 2025-12-04T09:45:19.5495489Z Running 0 items in this shard: 2025-12-04T09:45:19.5495664Z 2025-12-04T09:45:19.5496072Z Finished inductor/test_torchinductor_codegen_dynamic_shapes 2/4 ... [2025-12-04 09:45:19.549118][2225344.006438539], took 0.18min 2025-12-04T09:45:19.5499237Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:45:19.5610069Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:45:19.5611868Z Running inductor/test_torchinductor_opinfo 3/12 ... [2025-12-04 09:45:19.561000][2225344.018322953] 2025-12-04T09:45:19.5612233Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:45:19.5613483Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '--shard-id=3', '--num-shards=12', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:45:19.561213] 2025-12-04T09:45:32.7262187Z 2025-12-04T09:45:32.7263345Z inductor/test_torchinductor_opinfo 3/12 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_3.12_d77c95fa646ca8a9_.log 2025-12-04T09:45:32.7283713Z Running 50 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_interpolate_bilinear_cuda_float64 2025-12-04T09:45:32.7294705Z 2025-12-04T09:45:32.7294845Z Finished inductor/test_torchinductor_opinfo 3/12 ... [2025-12-04 09:45:32.725947][2225357.183266817], took 0.22min 2025-12-04T09:45:32.7295253Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:45:32.7378689Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:45:32.7384145Z Running inductor/test_torchinductor_opinfo 9/12 ... [2025-12-04 09:45:32.737672][2225357.194996793] 2025-12-04T09:45:32.7384395Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:45:32.7384869Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_torchinductor_opinfo.py', '--shard-id=9', '--num-shards=12', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:45:32.737875] 2025-12-04T09:47:27.9212237Z 2025-12-04T09:47:27.9213143Z inductor/test_torchinductor_opinfo 9/12 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_torchinductor_opinfo_9.12_79aaa6eac1ac94be_.log 2025-12-04T09:47:27.9232729Z Running 100 items in this shard: test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_nn_functional_adaptive_avg_pool2d_cuda_float16, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32, test/inductor/test_torchinductor_opinfo.py::TestInductorOpInfoCUDA::test_comprehensive_stft_cuda_float32 2025-12-04T09:47:27.9251469Z 2025-12-04T09:47:27.9251634Z Finished inductor/test_torchinductor_opinfo 9/12 ... [2025-12-04 09:47:27.921038][2225472.378358712], took 1.92min 2025-12-04T09:47:27.9252085Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:47:27.9334286Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:47:27.9336199Z Running dynamo/test_dynamic_shapes 1/2 ... [2025-12-04 09:47:27.933530][2225472.390854266] 2025-12-04T09:47:27.9336416Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:47:27.9338608Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_dynamic_shapes.py', '--shard-id=1', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:47:27.933746] 2025-12-04T09:47:37.2125389Z 2025-12-04T09:47:37.2126528Z dynamo/test_dynamic_shapes 1/2 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_dynamic_shapes_1.2_08b7077c23812445_.log 2025-12-04T09:47:37.2127026Z Running 0 items in this shard: 2025-12-04T09:47:37.2127112Z 2025-12-04T09:47:37.2127243Z Finished dynamo/test_dynamic_shapes 1/2 ... [2025-12-04 09:47:37.212264][2225481.669584018], took 0.15min 2025-12-04T09:47:37.2129799Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:47:37.2237972Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:47:37.2238617Z Running inductor/test_cpu_repro 5/5 ... [2025-12-04 09:47:37.223755][2225481.681078587] 2025-12-04T09:47:37.2239067Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:47:37.2241342Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_cpu_repro.py', '--shard-id=5', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:47:37.223954] 2025-12-04T09:47:45.4625993Z 2025-12-04T09:47:45.4626895Z inductor/test_cpu_repro 5/5 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cpu_repro_5.5_e8ed1312cf9a25ae_.log 2025-12-04T09:47:45.4627215Z Running 0 items in this shard: 2025-12-04T09:47:45.4627297Z 2025-12-04T09:47:45.4627416Z Finished inductor/test_cpu_repro 5/5 ... [2025-12-04 09:47:45.462273][2225489.919594407], took 0.14min 2025-12-04T09:47:45.4628551Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:47:45.4738520Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:47:45.4740493Z Running inductor/test_inductor_freezing 1/1 ... [2025-12-04 09:47:45.473940][2225489.931261163] 2025-12-04T09:47:45.4740743Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:47:45.4742845Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_inductor_freezing.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:47:45.474189] 2025-12-04T09:49:26.2102016Z 2025-12-04T09:49:26.2102983Z PRINTING LOG FILE of inductor/test_inductor_freezing 1/1 (test/test-reports/inductor.test_inductor_freezing_1.1_b201467f2f95967c_.log) 2025-12-04T09:49:26.2104499Z Test results will be stored in test-reports/python-pytest/inductor.test_inductor_freezing/inductor.test_inductor_freezing-82f27a0beef4242a.xml 2025-12-04T09:49:26.2104886Z ============================= test session starts ============================== 2025-12-04T09:49:26.2105122Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:49:26.2105371Z cachedir: .pytest_cache 2025-12-04T09:49:26.2105608Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:49:26.2105874Z rootdir: /var/lib/jenkins/pytorch 2025-12-04T09:49:26.2106003Z configfile: pytest.ini 2025-12-04T09:49:26.2106240Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:49:26.2106489Z collecting ... collected 48 items 2025-12-04T09:49:26.2106642Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:49:26.2118978Z Running 100 items in this shard: test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu, test/inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu 2025-12-04T09:49:26.2130843Z 2025-12-04T09:49:26.2131037Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [4.7162s] [ 1%] 2025-12-04T09:49:26.2131519Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2132000Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.4033s] [ 2%] 2025-12-04T09:49:26.2132410Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.1367s] [ 2%] 2025-12-04T09:49:26.2132808Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.3047s] [ 2%] 2025-12-04T09:49:26.2133207Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.3065s] [ 2%] 2025-12-04T09:49:26.2133606Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.8926s] [ 2%] 2025-12-04T09:49:26.2134006Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.7583s] [ 2%] 2025-12-04T09:49:26.2134407Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.9130s] [ 2%] 2025-12-04T09:49:26.2134802Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.3642s] [ 2%] 2025-12-04T09:49:26.2135199Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.4604s] [ 2%] 2025-12-04T09:49:26.2135598Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.3937s] [ 2%] 2025-12-04T09:49:26.2136034Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [0.8786s] [ 2%] 2025-12-04T09:49:26.2136432Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.0473s] [ 2%] 2025-12-04T09:49:26.2136870Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.8847s] [ 2%] 2025-12-04T09:49:26.2137265Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.4815s] [ 2%] 2025-12-04T09:49:26.2137667Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.3913s] [ 2%] 2025-12-04T09:49:26.2138063Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.6226s] [ 2%] 2025-12-04T09:49:26.2138462Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.6011s] [ 2%] 2025-12-04T09:49:26.2138863Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [0.8914s] [ 2%] 2025-12-04T09:49:26.2139257Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.3420s] [ 2%] 2025-12-04T09:49:26.2139653Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [0.9698s] [ 2%] 2025-12-04T09:49:26.2140088Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.6286s] [ 2%] 2025-12-04T09:49:26.2140491Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.8100s] [ 2%] 2025-12-04T09:49:26.2140890Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.7601s] [ 2%] 2025-12-04T09:49:26.2141289Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.1848s] [ 2%] 2025-12-04T09:49:26.2141687Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.3902s] [ 2%] 2025-12-04T09:49:26.2142087Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.6944s] [ 2%] 2025-12-04T09:49:26.2142491Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.2763s] [ 2%] 2025-12-04T09:49:26.2142891Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.2083s] [ 2%] 2025-12-04T09:49:26.2143288Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.2002s] [ 2%] 2025-12-04T09:49:26.2143683Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.0791s] [ 2%] 2025-12-04T09:49:26.2144078Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.2334s] [ 2%] 2025-12-04T09:49:26.2144475Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.9387s] [ 2%] 2025-12-04T09:49:26.2144875Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.6752s] [ 2%] 2025-12-04T09:49:26.2145329Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.9322s] [ 2%] 2025-12-04T09:49:26.2145726Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.2119s] [ 2%] 2025-12-04T09:49:26.2146122Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.0943s] [ 2%] 2025-12-04T09:49:26.2146523Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.4846s] [ 2%] 2025-12-04T09:49:26.2146972Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.2102s] [ 2%] 2025-12-04T09:49:26.2147370Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.1972s] [ 2%] 2025-12-04T09:49:26.2147767Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.1349s] [ 2%] 2025-12-04T09:49:26.2148162Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.5422s] [ 2%] 2025-12-04T09:49:26.2148557Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.1583s] [ 2%] 2025-12-04T09:49:26.2148952Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [0.9261s] [ 2%] 2025-12-04T09:49:26.2149349Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.9690s] [ 2%] 2025-12-04T09:49:26.2149781Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.9531s] [ 2%] 2025-12-04T09:49:26.2150179Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [1.7575s] [ 2%] 2025-12-04T09:49:26.2150574Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [2.2597s] [ 2%] 2025-12-04T09:49:26.2150973Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [0.6902s] [ 2%] 2025-12-04T09:49:26.2151369Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu <- test/inductor/test_torchinductor.py FAILED [0.9460s] [ 2%] 2025-12-04T09:49:26.2151843Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2152381Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2152917Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2153450Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2153984Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2154517Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2155081Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0005s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2155615Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2156150Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2156683Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2157250Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2157779Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2158309Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2158864Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2159398Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2159935Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2160470Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2161004Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2161537Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2162075Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2162609Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2163144Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2163685Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2164255Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2164787Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2165322Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2165856Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2166392Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2166962Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2167498Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2168031Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2168606Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2169141Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2169673Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2170205Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2170737Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2171273Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2171847Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2172380Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2172920Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2173455Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2174019Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2174563Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2175095Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2175627Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2176162Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2176697Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2177283Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0003s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2177817Z inductor/test_inductor_freezing.py::FreezingCpuTests::test_cpp_wrapper_cpu SKIPPED [0.0004s] (Test is enabled but --rerun-disabled-tests verification mode is set, so only disabled tests are run) [ 2%] 2025-12-04T09:49:26.2178105Z 2025-12-04T09:49:26.2178195Z =================================== FAILURES =================================== 2025-12-04T09:49:26.2178386Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2178557Z Traceback (most recent call last): 2025-12-04T09:49:26.2178764Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2178962Z return value(self) 2025-12-04T09:49:26.2179173Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2179682Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2179916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2180153Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2180408Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2180680Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2180847Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2180935Z 2025-12-04T09:49:26.2180990Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2181197Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2181448Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2181583Z 2025-12-04T09:49:26.2181661Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2181923Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2182110Z 2025-12-04T09:49:26.2182201Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2182404Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2182591Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2182762Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2183095Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2183443Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2183593Z graph_break [] 2025-12-04T09:49:26.2183736Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2183906Z Traceback (most recent call last): 2025-12-04T09:49:26.2184106Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2184301Z return value(self) 2025-12-04T09:49:26.2184508Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2184741Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2184970Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2185204Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2185456Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2185723Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2185884Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2185973Z 2025-12-04T09:49:26.2186027Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2186230Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2186479Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2186614Z 2025-12-04T09:49:26.2186719Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2187027Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2187211Z 2025-12-04T09:49:26.2187300Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2187497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2187654Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2187786Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2188114Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2188457Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2188605Z graph_break [] 2025-12-04T09:49:26.2189611Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2189773Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2189901Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2190066Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2190408Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2190713Z graph_break [] 2025-12-04T09:49:26.2190853Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2191021Z Traceback (most recent call last): 2025-12-04T09:49:26.2191217Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2191412Z return value(self) 2025-12-04T09:49:26.2191619Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2191885Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2192110Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2192343Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2192595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2192861Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2193022Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2193110Z 2025-12-04T09:49:26.2193163Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2193367Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2193619Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2193753Z 2025-12-04T09:49:26.2193834Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2194094Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2194279Z 2025-12-04T09:49:26.2194368Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2194566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2194722Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2194852Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2195182Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2195523Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2195704Z graph_break [] 2025-12-04T09:49:26.2195834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2195993Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2196118Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2196286Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2196624Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2196985Z graph_break [] 2025-12-04T09:49:26.2197112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2197267Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2197394Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2197558Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2197900Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2198294Z graph_break [] 2025-12-04T09:49:26.2198434Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2198602Z Traceback (most recent call last): 2025-12-04T09:49:26.2198799Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2198991Z return value(self) 2025-12-04T09:49:26.2199195Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2199425Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2199646Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2199881Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2200178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2200447Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2200608Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2200703Z 2025-12-04T09:49:26.2200757Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2200965Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2201220Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2201360Z 2025-12-04T09:49:26.2201436Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2201700Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2201892Z 2025-12-04T09:49:26.2201983Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2202191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2202353Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2202490Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2202822Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2203172Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2203326Z graph_break [] 2025-12-04T09:49:26.2203462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2203623Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2203756Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2203969Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2204320Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2204630Z graph_break [] 2025-12-04T09:49:26.2204766Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2204929Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2205060Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2205230Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2205579Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2205889Z graph_break [] 2025-12-04T09:49:26.2206026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2206191Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2206324Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2206494Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2206868Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2207177Z graph_break [] 2025-12-04T09:49:26.2207325Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2207501Z Traceback (most recent call last): 2025-12-04T09:49:26.2207706Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2207911Z return value(self) 2025-12-04T09:49:26.2208126Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2208409Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2208636Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2208875Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2209133Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2209406Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2209573Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2209661Z 2025-12-04T09:49:26.2209718Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2209927Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2210185Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2210320Z 2025-12-04T09:49:26.2210406Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2210671Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2210858Z 2025-12-04T09:49:26.2210950Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2211155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2211315Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2211452Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2211782Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2212132Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2212324Z graph_break [] 2025-12-04T09:49:26.2212466Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2212627Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2212762Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2212928Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2213274Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2213589Z graph_break [] 2025-12-04T09:49:26.2213719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2213881Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2214014Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2214185Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2214529Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2214841Z graph_break [] 2025-12-04T09:49:26.2214976Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2215137Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2215270Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2215439Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2215786Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2216094Z graph_break [] 2025-12-04T09:49:26.2216232Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2216394Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2216555Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2216727Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2217111Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2217424Z graph_break [] 2025-12-04T09:49:26.2217571Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2217747Z Traceback (most recent call last): 2025-12-04T09:49:26.2217952Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2218152Z return value(self) 2025-12-04T09:49:26.2218366Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2218603Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2218836Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2219073Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2219337Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2219610Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2219777Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2219870Z 2025-12-04T09:49:26.2219923Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2220134Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2220389Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2220529Z 2025-12-04T09:49:26.2220643Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2220914Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2221107Z 2025-12-04T09:49:26.2221197Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2221403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2221568Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2221704Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2222036Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2222386Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2222539Z graph_break [] 2025-12-04T09:49:26.2222676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2222840Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2222975Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2223145Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2223492Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2223808Z graph_break [] 2025-12-04T09:49:26.2223943Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2224106Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2224239Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2224411Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2224765Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2225111Z graph_break [] 2025-12-04T09:49:26.2225247Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2225412Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2225544Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2225715Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2226060Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2226371Z graph_break [] 2025-12-04T09:49:26.2226508Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2226670Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2226833Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2227007Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2227350Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2227655Z graph_break [] 2025-12-04T09:49:26.2227784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2227941Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2228069Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2228240Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2228618Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2228935Z graph_break [] 2025-12-04T09:49:26.2229081Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2229262Z Traceback (most recent call last): 2025-12-04T09:49:26.2229465Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2229666Z return value(self) 2025-12-04T09:49:26.2229880Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2230119Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2230349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2230587Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2230847Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2231115Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2231277Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2231367Z 2025-12-04T09:49:26.2231420Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2231622Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2231870Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2232007Z 2025-12-04T09:49:26.2232080Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2232339Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2232529Z 2025-12-04T09:49:26.2232616Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2232818Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2232976Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2233139Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2233466Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2233808Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2233956Z graph_break [] 2025-12-04T09:49:26.2234084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2234243Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2234370Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2234536Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2234878Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2235189Z graph_break [] 2025-12-04T09:49:26.2235317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2235474Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2235600Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2235765Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2236103Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2236414Z graph_break [] 2025-12-04T09:49:26.2236545Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2236704Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2236929Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2237095Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2237434Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2237742Z graph_break [] 2025-12-04T09:49:26.2237869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2238024Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2238150Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2238312Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2238649Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2238957Z graph_break [] 2025-12-04T09:49:26.2239085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2239243Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2239368Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2239530Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2239871Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2240174Z graph_break [] 2025-12-04T09:49:26.2240301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2240456Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2240581Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2240743Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2241089Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2241438Z graph_break [] 2025-12-04T09:49:26.2241577Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2241746Z Traceback (most recent call last): 2025-12-04T09:49:26.2241943Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2242137Z return value(self) 2025-12-04T09:49:26.2242344Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2242574Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2242794Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2243029Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2243285Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2243551Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2243711Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2243801Z 2025-12-04T09:49:26.2243852Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2244055Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2244303Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2244440Z 2025-12-04T09:49:26.2244513Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2244770Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2244957Z 2025-12-04T09:49:26.2245108Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2245309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2245463Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2245591Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2245918Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2246263Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2246410Z graph_break [] 2025-12-04T09:49:26.2246540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2246694Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2246874Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2247041Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2247384Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2247690Z graph_break [] 2025-12-04T09:49:26.2247818Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2247976Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2248103Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2248265Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2248605Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2248909Z graph_break [] 2025-12-04T09:49:26.2249041Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2249241Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2249367Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2249528Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2249871Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2250171Z graph_break [] 2025-12-04T09:49:26.2250298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2250452Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2250577Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2250741Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2251079Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2251386Z graph_break [] 2025-12-04T09:49:26.2251514Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2251668Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2251792Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2251955Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2252290Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2252592Z graph_break [] 2025-12-04T09:49:26.2252719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2252873Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2253029Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2253198Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2253543Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2253846Z graph_break [] 2025-12-04T09:49:26.2253974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2254128Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2254253Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2254414Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2254756Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2255060Z graph_break [] 2025-12-04T09:49:26.2255201Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2255369Z Traceback (most recent call last): 2025-12-04T09:49:26.2269656Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2269886Z return value(self) 2025-12-04T09:49:26.2270108Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2270349Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2270594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2270832Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2271090Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2271366Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2271603Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2271695Z 2025-12-04T09:49:26.2271749Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2271957Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2272210Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2272349Z 2025-12-04T09:49:26.2272426Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2272692Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2272883Z 2025-12-04T09:49:26.2272975Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2273180Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2273339Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2273480Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2273814Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2274165Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2274318Z graph_break [] 2025-12-04T09:49:26.2274451Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2274608Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2274738Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2274904Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2275286Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2275603Z graph_break [] 2025-12-04T09:49:26.2275736Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2275893Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2276023Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2276190Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2276533Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2276959Z graph_break [] 2025-12-04T09:49:26.2277090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2277248Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2277380Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2277546Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2277893Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2278201Z graph_break [] 2025-12-04T09:49:26.2278332Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2278489Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2278618Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2278782Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2279120Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2279427Z graph_break [] 2025-12-04T09:49:26.2279557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2279762Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2279890Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2280055Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2280396Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2280699Z graph_break [] 2025-12-04T09:49:26.2280831Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2280988Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2281118Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2281286Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2281628Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2281939Z graph_break [] 2025-12-04T09:49:26.2282071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2282228Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2282356Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2282519Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2282857Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2283162Z graph_break [] 2025-12-04T09:49:26.2283295Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2283495Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2283634Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2283798Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2284137Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2284444Z graph_break [] 2025-12-04T09:49:26.2284587Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2284762Z Traceback (most recent call last): 2025-12-04T09:49:26.2284965Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2285164Z return value(self) 2025-12-04T09:49:26.2285374Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2285610Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2285839Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2286076Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2286332Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2286603Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2286793Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2286886Z 2025-12-04T09:49:26.2286940Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2287145Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2287397Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2287535Z 2025-12-04T09:49:26.2287615Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2287913Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2288105Z 2025-12-04T09:49:26.2288197Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2288395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2288553Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2288687Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2289012Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2289357Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2289508Z graph_break [] 2025-12-04T09:49:26.2289643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2289802Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2289935Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2290102Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2290444Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2290752Z graph_break [] 2025-12-04T09:49:26.2290884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2291045Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2291173Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2291338Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2291714Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2292030Z graph_break [] 2025-12-04T09:49:26.2292161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2292317Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2292441Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2292608Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2292947Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2293252Z graph_break [] 2025-12-04T09:49:26.2293382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2293539Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2293671Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2293840Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2294179Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2294487Z graph_break [] 2025-12-04T09:49:26.2294616Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2294766Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2294894Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2295061Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2295409Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2295719Z graph_break [] 2025-12-04T09:49:26.2295875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2296033Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2296163Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2296330Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2296675Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2297020Z graph_break [] 2025-12-04T09:49:26.2297156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2297315Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2297449Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2297619Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2297964Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2298280Z graph_break [] 2025-12-04T09:49:26.2298415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2298576Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2298708Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2298875Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2299219Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2299527Z graph_break [] 2025-12-04T09:49:26.2299692Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2299852Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2299988Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2300155Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2300495Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2300804Z graph_break [] 2025-12-04T09:49:26.2300950Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2301125Z Traceback (most recent call last): 2025-12-04T09:49:26.2301329Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2301528Z return value(self) 2025-12-04T09:49:26.2301745Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2301980Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2302211Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2302449Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2302706Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2302981Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2303148Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2303238Z 2025-12-04T09:49:26.2303294Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2303503Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2303757Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2303895Z 2025-12-04T09:49:26.2303974Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2304270Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2304457Z 2025-12-04T09:49:26.2304550Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2304753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2304915Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2305049Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2305378Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2305723Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2305875Z graph_break [] 2025-12-04T09:49:26.2306011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2306175Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2306303Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2306469Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2306849Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2307155Z graph_break [] 2025-12-04T09:49:26.2307284Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2307440Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2307568Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2307731Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2308110Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2308415Z graph_break [] 2025-12-04T09:49:26.2308544Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2308698Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2308828Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2308992Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2309334Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2309636Z graph_break [] 2025-12-04T09:49:26.2309764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2309922Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2310048Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2310216Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2310551Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2310855Z graph_break [] 2025-12-04T09:49:26.2310982Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2311135Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2311264Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2311430Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2311775Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2312141Z graph_break [] 2025-12-04T09:49:26.2312271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2312426Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2312555Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2312721Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2313056Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2313358Z graph_break [] 2025-12-04T09:49:26.2313485Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2313639Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2313765Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2313932Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2314270Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2314570Z graph_break [] 2025-12-04T09:49:26.2314698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2314852Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2314978Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2315141Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2315479Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2315780Z graph_break [] 2025-12-04T09:49:26.2315944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2316104Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2316228Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2316391Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2316727Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2317070Z graph_break [] 2025-12-04T09:49:26.2317197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2317349Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2317474Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2317634Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2317973Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2318275Z graph_break [] 2025-12-04T09:49:26.2318413Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2318579Z Traceback (most recent call last): 2025-12-04T09:49:26.2318774Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2318967Z return value(self) 2025-12-04T09:49:26.2319172Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2319403Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2319625Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2319856Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2320109Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2320416Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2320574Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2320662Z 2025-12-04T09:49:26.2320712Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2320914Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2321160Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2321295Z 2025-12-04T09:49:26.2321370Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2321626Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2321812Z 2025-12-04T09:49:26.2321899Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2322098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2322249Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2322375Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2322696Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2323034Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2323178Z graph_break [] 2025-12-04T09:49:26.2323304Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2323457Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2323580Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2323769Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2324107Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2324412Z graph_break [] 2025-12-04T09:49:26.2324538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2324688Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2324812Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2324973Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2325309Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2325610Z graph_break [] 2025-12-04T09:49:26.2325739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2325891Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2326017Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2326178Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2326513Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2326864Z graph_break [] 2025-12-04T09:49:26.2326989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2327140Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2327263Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2327424Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2327763Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2328106Z graph_break [] 2025-12-04T09:49:26.2328231Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2328381Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2328503Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2328663Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2328997Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2329300Z graph_break [] 2025-12-04T09:49:26.2329425Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2329575Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2329700Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2329860Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2330199Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2330500Z graph_break [] 2025-12-04T09:49:26.2330624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2330774Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2330895Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2331055Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2331388Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2331718Z graph_break [] 2025-12-04T09:49:26.2331845Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2331999Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2332122Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2332282Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2332616Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2332915Z graph_break [] 2025-12-04T09:49:26.2333039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2333190Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2333312Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2333472Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2333807Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2334108Z graph_break [] 2025-12-04T09:49:26.2334231Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2334382Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2334505Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2334664Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2334999Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2335302Z graph_break [] 2025-12-04T09:49:26.2335427Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2335580Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2335728Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2335886Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2336217Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2336515Z graph_break [] 2025-12-04T09:49:26.2336651Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2336857Z Traceback (most recent call last): 2025-12-04T09:49:26.2337050Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2337242Z return value(self) 2025-12-04T09:49:26.2337445Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2337673Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2337898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2338126Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2338376Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2338638Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2338794Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2338882Z 2025-12-04T09:49:26.2338933Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2339132Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2339377Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2339510Z 2025-12-04T09:49:26.2339629Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2339886Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2340069Z 2025-12-04T09:49:26.2340155Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2340348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2340498Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2340623Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2340942Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2341279Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2341422Z graph_break [] 2025-12-04T09:49:26.2341554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2341707Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2341831Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2341992Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2342327Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2342629Z graph_break [] 2025-12-04T09:49:26.2342754Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2342906Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2343032Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2343193Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2343532Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2345884Z graph_break [] 2025-12-04T09:49:26.2346013Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2346167Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2346290Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2346455Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2346922Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2347224Z graph_break [] 2025-12-04T09:49:26.2347350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2347503Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2347633Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2347798Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2348158Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2348460Z graph_break [] 2025-12-04T09:49:26.2348587Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2348742Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2348866Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2349028Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2349406Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2349708Z graph_break [] 2025-12-04T09:49:26.2349838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2349993Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2350118Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2350283Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2350619Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2350925Z graph_break [] 2025-12-04T09:49:26.2351052Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2351206Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2351331Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2351497Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2351834Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2352144Z graph_break [] 2025-12-04T09:49:26.2352273Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2352430Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2352554Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2352717Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2353055Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2353362Z graph_break [] 2025-12-04T09:49:26.2353492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2353646Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2353791Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2354013Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2354347Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2354650Z graph_break [] 2025-12-04T09:49:26.2354776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2354929Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2355053Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2355215Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2355553Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2355857Z graph_break [] 2025-12-04T09:49:26.2355984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2356136Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2356261Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2356422Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2356785Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2357089Z graph_break [] 2025-12-04T09:49:26.2357217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2357369Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2357540Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2357704Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2358042Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2358345Z graph_break [] 2025-12-04T09:49:26.2358486Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2358654Z Traceback (most recent call last): 2025-12-04T09:49:26.2358852Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2359044Z return value(self) 2025-12-04T09:49:26.2359250Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2359480Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2359709Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2359942Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2360194Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2360459Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2360617Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2360706Z 2025-12-04T09:49:26.2360759Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2360962Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2361212Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2361346Z 2025-12-04T09:49:26.2361423Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2361693Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2361894Z 2025-12-04T09:49:26.2361986Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2362217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2362375Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2362504Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2362833Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2363174Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2363324Z graph_break [] 2025-12-04T09:49:26.2363457Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2363616Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2363748Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2363921Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2364265Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2364573Z graph_break [] 2025-12-04T09:49:26.2364704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2364864Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2364998Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2365165Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2365529Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2365840Z graph_break [] 2025-12-04T09:49:26.2365972Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2366130Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2366260Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2366428Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2366804Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2367110Z graph_break [] 2025-12-04T09:49:26.2367242Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2367400Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2367527Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2367695Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2368036Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2368342Z graph_break [] 2025-12-04T09:49:26.2368472Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2368631Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2368759Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2368926Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2369265Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2369573Z graph_break [] 2025-12-04T09:49:26.2369708Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2369890Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2370020Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2370202Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2370543Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2370851Z graph_break [] 2025-12-04T09:49:26.2370980Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2371136Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2371262Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2371428Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2371772Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2372083Z graph_break [] 2025-12-04T09:49:26.2372213Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2372371Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2372499Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2372667Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2373005Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2373311Z graph_break [] 2025-12-04T09:49:26.2373442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2373599Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2373764Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2373933Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2374274Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2374581Z graph_break [] 2025-12-04T09:49:26.2374714Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2374871Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2375000Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2375167Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2375511Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2375815Z graph_break [] 2025-12-04T09:49:26.2375947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2376107Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2376239Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2376407Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2376801Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2377110Z graph_break [] 2025-12-04T09:49:26.2377243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2377400Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2377529Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2377701Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2378038Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2378386Z graph_break [] 2025-12-04T09:49:26.2378518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2378675Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2378805Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2378974Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2379315Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2379623Z graph_break [] 2025-12-04T09:49:26.2379768Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2379944Z Traceback (most recent call last): 2025-12-04T09:49:26.2380145Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2380346Z return value(self) 2025-12-04T09:49:26.2380559Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2380794Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2381019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2381257Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2381513Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2381783Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2381972Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2382060Z 2025-12-04T09:49:26.2382113Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2382314Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2382566Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2382701Z 2025-12-04T09:49:26.2382780Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2383042Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2383227Z 2025-12-04T09:49:26.2383319Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2383519Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2383676Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2383812Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2384141Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2384486Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2384637Z graph_break [] 2025-12-04T09:49:26.2384768Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2384925Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2385055Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2385222Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2385565Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2385875Z graph_break [] 2025-12-04T09:49:26.2386006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2386181Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2386324Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2386491Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2386870Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2387179Z graph_break [] 2025-12-04T09:49:26.2387310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2387465Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2387597Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2387765Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2388110Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2388426Z graph_break [] 2025-12-04T09:49:26.2388557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2388717Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2388847Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2388926Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2389158Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2389200Z graph_break [] 2025-12-04T09:49:26.2389276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2389353Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2389411Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2389486Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2389715Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2389754Z graph_break [] 2025-12-04T09:49:26.2389829Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2389873Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2389928Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2390001Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2390233Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2390271Z graph_break [] 2025-12-04T09:49:26.2390345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2390387Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2390444Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2390515Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2390743Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2390780Z graph_break [] 2025-12-04T09:49:26.2390854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2390895Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2390951Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2391026Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2391275Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2391326Z graph_break [] 2025-12-04T09:49:26.2391401Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2391442Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2391500Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2391571Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2391800Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2391836Z graph_break [] 2025-12-04T09:49:26.2391912Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2391956Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2392012Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2392086Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2392315Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2392351Z graph_break [] 2025-12-04T09:49:26.2392425Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2392466Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2392523Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2392595Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2392843Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2392882Z graph_break [] 2025-12-04T09:49:26.2392956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2392997Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2393054Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2393126Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2393355Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2393391Z graph_break [] 2025-12-04T09:49:26.2393466Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2393509Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2393566Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2393640Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2393868Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2393906Z graph_break [] 2025-12-04T09:49:26.2393978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2394020Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2394075Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2394148Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2394376Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2394427Z graph_break [] 2025-12-04T09:49:26.2394513Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2394573Z Traceback (most recent call last): 2025-12-04T09:49:26.2394694Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2394735Z return value(self) 2025-12-04T09:49:26.2394880Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2394932Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2395069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2395130Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2395290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2395365Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2395418Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2395421Z 2025-12-04T09:49:26.2395474Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2395593Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2395692Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2395694Z 2025-12-04T09:49:26.2395767Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2395920Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2395922Z 2025-12-04T09:49:26.2396008Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2396112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2396155Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2396215Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2396445Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2396520Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2396557Z graph_break [] 2025-12-04T09:49:26.2396632Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2396673Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2396731Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2396861Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2397093Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2397134Z graph_break [] 2025-12-04T09:49:26.2397208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2397252Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2397308Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2397382Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2397610Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2397649Z graph_break [] 2025-12-04T09:49:26.2397722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2397766Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2397822Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2397895Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2398142Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2398195Z graph_break [] 2025-12-04T09:49:26.2398267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2398311Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2398366Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2398439Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2398665Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2398705Z graph_break [] 2025-12-04T09:49:26.2398779Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2398823Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2398878Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2398951Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2399177Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2399217Z graph_break [] 2025-12-04T09:49:26.2399290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2399333Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2399388Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2399461Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2399713Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2399755Z graph_break [] 2025-12-04T09:49:26.2399828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2399871Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2399925Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2399999Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2400226Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2400265Z graph_break [] 2025-12-04T09:49:26.2400337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2400382Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2400437Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2400510Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2400741Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2400777Z graph_break [] 2025-12-04T09:49:26.2400852Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2400893Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2400948Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2401020Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2401251Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2401299Z graph_break [] 2025-12-04T09:49:26.2401373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2401426Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2401484Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2401555Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2401785Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2401822Z graph_break [] 2025-12-04T09:49:26.2401895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2401936Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2401992Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2402065Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2402293Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2402331Z graph_break [] 2025-12-04T09:49:26.2402405Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2402446Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2402502Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2402574Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2402801Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2402837Z graph_break [] 2025-12-04T09:49:26.2402931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2402974Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2403032Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2403103Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2403335Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2403372Z graph_break [] 2025-12-04T09:49:26.2403447Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2403488Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2403544Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2403615Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2403845Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2403883Z graph_break [] 2025-12-04T09:49:26.2403957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2403999Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2404055Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2404126Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2404354Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2404392Z graph_break [] 2025-12-04T09:49:26.2404477Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2404527Z Traceback (most recent call last): 2025-12-04T09:49:26.2404647Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2404705Z return value(self) 2025-12-04T09:49:26.2404865Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2404917Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2405054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2405115Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2405275Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2405349Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2405401Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2405404Z 2025-12-04T09:49:26.2405459Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2405578Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2405678Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2405680Z 2025-12-04T09:49:26.2405753Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2405904Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2405906Z 2025-12-04T09:49:26.2405992Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2406067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2406110Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2406167Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2406414Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2406491Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2406528Z graph_break [] 2025-12-04T09:49:26.2406602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2406643Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2406700Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2406799Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2407031Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2407068Z graph_break [] 2025-12-04T09:49:26.2407146Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2407188Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2407245Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2407318Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2407551Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2407591Z graph_break [] 2025-12-04T09:49:26.2407664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2407707Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2407762Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2407835Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2408064Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2408124Z graph_break [] 2025-12-04T09:49:26.2408218Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2408261Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2408315Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2408388Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2408616Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2408656Z graph_break [] 2025-12-04T09:49:26.2408729Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2408772Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2408828Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2408902Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2409130Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2409169Z graph_break [] 2025-12-04T09:49:26.2409243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2409286Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2409340Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2409413Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2409639Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2409708Z graph_break [] 2025-12-04T09:49:26.2409782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2409826Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2409881Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2409953Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2410181Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2410219Z graph_break [] 2025-12-04T09:49:26.2410291Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2410334Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2410388Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2410461Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2410691Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2414379Z graph_break [] 2025-12-04T09:49:26.2414462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2414506Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2414562Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2414638Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2414869Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2414909Z graph_break [] 2025-12-04T09:49:26.2414989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2415034Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2415120Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2415193Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2415439Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2415477Z graph_break [] 2025-12-04T09:49:26.2415553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2415595Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2415653Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2415725Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2415957Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2415997Z graph_break [] 2025-12-04T09:49:26.2416073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2416115Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2416173Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2416244Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2416475Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2416511Z graph_break [] 2025-12-04T09:49:26.2416588Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2416629Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2416708Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2416820Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2417053Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2417091Z graph_break [] 2025-12-04T09:49:26.2417165Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2417206Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2417261Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2417331Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2417559Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2417595Z graph_break [] 2025-12-04T09:49:26.2417671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2417715Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2417772Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2417843Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2418074Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2418112Z graph_break [] 2025-12-04T09:49:26.2418187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2418229Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2418285Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2418358Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2418588Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2418668Z graph_break [] 2025-12-04T09:49:26.2418757Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2418804Z Traceback (most recent call last): 2025-12-04T09:49:26.2418928Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2418968Z return value(self) 2025-12-04T09:49:26.2419115Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2419171Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2419312Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2419376Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2419536Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2419614Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2419667Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2419670Z 2025-12-04T09:49:26.2419725Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2419844Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2419943Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2419945Z 2025-12-04T09:49:26.2420020Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2420173Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2420175Z 2025-12-04T09:49:26.2420288Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2420367Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2420411Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2420467Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2420699Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2420776Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2420814Z graph_break [] 2025-12-04T09:49:26.2420891Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2420933Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2420991Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2421066Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2421302Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2421341Z graph_break [] 2025-12-04T09:49:26.2421415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2421457Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2421514Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2421587Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2421815Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2421853Z graph_break [] 2025-12-04T09:49:26.2421927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2421982Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2422036Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2422122Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2422348Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2422387Z graph_break [] 2025-12-04T09:49:26.2422459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2422503Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2422558Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2422631Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2422859Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2422897Z graph_break [] 2025-12-04T09:49:26.2422972Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2423015Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2423071Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2423147Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2423376Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2423414Z graph_break [] 2025-12-04T09:49:26.2423486Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2423530Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2423605Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2423679Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2423908Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2423947Z graph_break [] 2025-12-04T09:49:26.2424019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2424063Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2424117Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2424189Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2424418Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2424458Z graph_break [] 2025-12-04T09:49:26.2424531Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2424575Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2424628Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2424703Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2424930Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2424968Z graph_break [] 2025-12-04T09:49:26.2425042Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2425085Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2425140Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2425215Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2425441Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2425511Z graph_break [] 2025-12-04T09:49:26.2425585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2425629Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2425684Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2425757Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2425989Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2426027Z graph_break [] 2025-12-04T09:49:26.2426105Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2426145Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2426202Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2426275Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2426503Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2426539Z graph_break [] 2025-12-04T09:49:26.2426615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2426657Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2426715Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2426821Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2427082Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2427122Z graph_break [] 2025-12-04T09:49:26.2427199Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2427241Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2427298Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2427370Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2427598Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2427634Z graph_break [] 2025-12-04T09:49:26.2427709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2427750Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2427808Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2427880Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2428111Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2428148Z graph_break [] 2025-12-04T09:49:26.2428224Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2428264Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2428319Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2428392Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2428621Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2428660Z graph_break [] 2025-12-04T09:49:26.2428735Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2428795Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2428867Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2428938Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2429173Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2429210Z graph_break [] 2025-12-04T09:49:26.2429283Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2429326Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2429383Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2429454Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2429685Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2429728Z graph_break [] 2025-12-04T09:49:26.2429814Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2429864Z Traceback (most recent call last): 2025-12-04T09:49:26.2429985Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2430026Z return value(self) 2025-12-04T09:49:26.2430171Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2430224Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2430361Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2430446Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2430607Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2430684Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2430739Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2430741Z 2025-12-04T09:49:26.2430797Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2430914Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2431014Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2431017Z 2025-12-04T09:49:26.2431090Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2431243Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2431245Z 2025-12-04T09:49:26.2431334Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2431410Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2431453Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2431511Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2431742Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2431819Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2431855Z graph_break [] 2025-12-04T09:49:26.2431932Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2431974Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2432033Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2432106Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2432350Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2432406Z graph_break [] 2025-12-04T09:49:26.2432477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2432519Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2432573Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2432645Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2432873Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2432913Z graph_break [] 2025-12-04T09:49:26.2432989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2433034Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2433089Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2433163Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2433391Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2433429Z graph_break [] 2025-12-04T09:49:26.2433502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2433545Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2433599Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2433672Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2433919Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2433964Z graph_break [] 2025-12-04T09:49:26.2434038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2434081Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2434137Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2434211Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2434438Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2434476Z graph_break [] 2025-12-04T09:49:26.2434548Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2434589Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2434644Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2434717Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2434946Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2434982Z graph_break [] 2025-12-04T09:49:26.2435054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2435096Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2435150Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2435222Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2435451Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2435488Z graph_break [] 2025-12-04T09:49:26.2435573Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2435627Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2435682Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2435753Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2435979Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2436018Z graph_break [] 2025-12-04T09:49:26.2436091Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2436138Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2436192Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2436268Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2436500Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2436541Z graph_break [] 2025-12-04T09:49:26.2436620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2436661Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2436717Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2436820Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2437050Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2437088Z graph_break [] 2025-12-04T09:49:26.2437196Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2437240Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2437301Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2437373Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2437603Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2437640Z graph_break [] 2025-12-04T09:49:26.2437716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2437757Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2437814Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2437885Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2438113Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2438150Z graph_break [] 2025-12-04T09:49:26.2438227Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2438267Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2438324Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2438396Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2438626Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2438663Z graph_break [] 2025-12-04T09:49:26.2438739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2438780Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2438839Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2438926Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2439162Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2439217Z graph_break [] 2025-12-04T09:49:26.2439293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2439335Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2439393Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2439467Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2439701Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2439739Z graph_break [] 2025-12-04T09:49:26.2439816Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2439861Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2439919Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2439992Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2440223Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2440259Z graph_break [] 2025-12-04T09:49:26.2440336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2440378Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2440435Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2440507Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2440761Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2440801Z graph_break [] 2025-12-04T09:49:26.2440876Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2440918Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2440973Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2441045Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2441276Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2441313Z graph_break [] 2025-12-04T09:49:26.2441401Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2441449Z Traceback (most recent call last): 2025-12-04T09:49:26.2441572Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2441612Z return value(self) 2025-12-04T09:49:26.2441759Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2441811Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2441951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2442013Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2442176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2442249Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2442305Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2442309Z 2025-12-04T09:49:26.2442362Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2442498Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2442606Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2442608Z 2025-12-04T09:49:26.2442683Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2442833Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2442838Z 2025-12-04T09:49:26.2442925Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2443002Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2443046Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2443104Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2443335Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2443412Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2443449Z graph_break [] 2025-12-04T09:49:26.2443526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2443570Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2443629Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2443700Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2443931Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2443969Z graph_break [] 2025-12-04T09:49:26.2444062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2444106Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2444160Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2444234Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2444460Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2444497Z graph_break [] 2025-12-04T09:49:26.2444569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2444611Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2444664Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2444736Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2444965Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2445004Z graph_break [] 2025-12-04T09:49:26.2445076Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2445118Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2445172Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2445244Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2445471Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2445506Z graph_break [] 2025-12-04T09:49:26.2445579Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2445619Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2445675Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2445759Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2445998Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2446033Z graph_break [] 2025-12-04T09:49:26.2446108Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2446148Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2446203Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2446274Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2446504Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2446540Z graph_break [] 2025-12-04T09:49:26.2446615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2446657Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2446711Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2446820Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2447050Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2447086Z graph_break [] 2025-12-04T09:49:26.2447159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2447199Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2447253Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2447348Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2447577Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2447615Z graph_break [] 2025-12-04T09:49:26.2447687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2447728Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2447783Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2447854Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2448082Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2448118Z graph_break [] 2025-12-04T09:49:26.2448192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2448233Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2448288Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2448360Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2448590Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2448626Z graph_break [] 2025-12-04T09:49:26.2448701Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2448742Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2448796Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2448867Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2449099Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2449155Z graph_break [] 2025-12-04T09:49:26.2449241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2449283Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2449337Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2449408Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2449634Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2449671Z graph_break [] 2025-12-04T09:49:26.2449743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2449785Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2449839Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2449912Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2450138Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2450176Z graph_break [] 2025-12-04T09:49:26.2450248Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2450289Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2450342Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2450413Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2450639Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2450708Z graph_break [] 2025-12-04T09:49:26.2450782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2450825Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2450879Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2450951Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2451177Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2451215Z graph_break [] 2025-12-04T09:49:26.2451286Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2451328Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2451381Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2451455Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2451681Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2451720Z graph_break [] 2025-12-04T09:49:26.2451792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2451833Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2451886Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2451958Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2452184Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2452221Z graph_break [] 2025-12-04T09:49:26.2452295Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2452338Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2452411Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2452496Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2452725Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2452761Z graph_break [] 2025-12-04T09:49:26.2452835Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2452876Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2452931Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2453002Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2453232Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2453269Z graph_break [] 2025-12-04T09:49:26.2453355Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2453400Z Traceback (most recent call last): 2025-12-04T09:49:26.2453520Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2453559Z return value(self) 2025-12-04T09:49:26.2453705Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2453756Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2453893Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2453953Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2454137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2454212Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2454266Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2454268Z 2025-12-04T09:49:26.2454318Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2454437Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2454534Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2454536Z 2025-12-04T09:49:26.2454610Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2454759Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2454762Z 2025-12-04T09:49:26.2454849Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2454924Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2454968Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2455025Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2455254Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2455328Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2455364Z graph_break [] 2025-12-04T09:49:26.2455437Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2455480Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2455534Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2455606Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2455835Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2455896Z graph_break [] 2025-12-04T09:49:26.2455970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2456011Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2456067Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2456138Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2456367Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2456403Z graph_break [] 2025-12-04T09:49:26.2456477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2456520Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2456576Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2456649Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2456916Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2456953Z graph_break [] 2025-12-04T09:49:26.2457026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2457068Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2457124Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2457195Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2457456Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2457494Z graph_break [] 2025-12-04T09:49:26.2457567Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2457609Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2457664Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2457734Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2457962Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2457998Z graph_break [] 2025-12-04T09:49:26.2458071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2458112Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2458167Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2458238Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2458467Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2458504Z graph_break [] 2025-12-04T09:49:26.2458578Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2458618Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2458673Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2458744Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2458974Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2459010Z graph_break [] 2025-12-04T09:49:26.2459085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2459150Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2459204Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2459288Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2459515Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2459553Z graph_break [] 2025-12-04T09:49:26.2459625Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2459666Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2459719Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2459790Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2460019Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2460058Z graph_break [] 2025-12-04T09:49:26.2460130Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2460171Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2460224Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2460296Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2460522Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2460559Z graph_break [] 2025-12-04T09:49:26.2460631Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2460674Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2460750Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2460825Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2461054Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2461091Z graph_break [] 2025-12-04T09:49:26.2461163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2461205Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2461259Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2461331Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2461558Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2461595Z graph_break [] 2025-12-04T09:49:26.2461668Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2461711Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2461764Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2461835Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2462061Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2462098Z graph_break [] 2025-12-04T09:49:26.2462170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2462211Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2462264Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2462337Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2462563Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2462629Z graph_break [] 2025-12-04T09:49:26.2462703Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2462745Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2462799Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2462871Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2463100Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2463136Z graph_break [] 2025-12-04T09:49:26.2463211Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2463252Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2463308Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2463380Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2463609Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2463645Z graph_break [] 2025-12-04T09:49:26.2463718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2463759Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2463814Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2463885Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2464137Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2464175Z graph_break [] 2025-12-04T09:49:26.2464250Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2464291Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2464346Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2464416Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2464646Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2464682Z graph_break [] 2025-12-04T09:49:26.2464755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2464795Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2464851Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2464922Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2465151Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2465188Z graph_break [] 2025-12-04T09:49:26.2465261Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2465301Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2465355Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2465426Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2465652Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2465690Z graph_break [] 2025-12-04T09:49:26.2465776Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2465834Z Traceback (most recent call last): 2025-12-04T09:49:26.2465964Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2466002Z return value(self) 2025-12-04T09:49:26.2466147Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2466198Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2466335Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2466395Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2466556Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2466630Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2466683Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2466686Z 2025-12-04T09:49:26.2466737Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2466891Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2466989Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2466992Z 2025-12-04T09:49:26.2467064Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2467214Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2467217Z 2025-12-04T09:49:26.2467302Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2467376Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2467457Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2467514Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2467743Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2467817Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2467853Z graph_break [] 2025-12-04T09:49:26.2467927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2467968Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2468023Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2468094Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2468323Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2468360Z graph_break [] 2025-12-04T09:49:26.2468435Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2468476Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2468533Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2468604Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2468834Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2468870Z graph_break [] 2025-12-04T09:49:26.2468944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2468985Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2469041Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2469114Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2469363Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2469412Z graph_break [] 2025-12-04T09:49:26.2469486Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2469526Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2469581Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2469653Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2469881Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2469920Z graph_break [] 2025-12-04T09:49:26.2469994Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2470037Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2470091Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2470164Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2470390Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2470427Z graph_break [] 2025-12-04T09:49:26.2470499Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2470541Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2470595Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2470668Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2470918Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2470958Z graph_break [] 2025-12-04T09:49:26.2471030Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2471072Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2471127Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2471201Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2471428Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2471465Z graph_break [] 2025-12-04T09:49:26.2471538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2471580Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2471635Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2471708Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2471937Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2471975Z graph_break [] 2025-12-04T09:49:26.2472047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2472089Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2472143Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2472216Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2472443Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2472480Z graph_break [] 2025-12-04T09:49:26.2472572Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2472623Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2472677Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2472750Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2472975Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2473012Z graph_break [] 2025-12-04T09:49:26.2473084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2473127Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2473184Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2473257Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2473485Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2473523Z graph_break [] 2025-12-04T09:49:26.2473597Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2473638Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2473693Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2473763Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2473991Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2474027Z graph_break [] 2025-12-04T09:49:26.2474120Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2474163Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2474218Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2474291Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2474518Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2474554Z graph_break [] 2025-12-04T09:49:26.2474628Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2474669Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2474724Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2474795Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2475023Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2475060Z graph_break [] 2025-12-04T09:49:26.2475135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2475175Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2475230Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2475301Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2475528Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2475565Z graph_break [] 2025-12-04T09:49:26.2475639Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2475680Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2475737Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2475826Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2476054Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2476102Z graph_break [] 2025-12-04T09:49:26.2476176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2476217Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2476273Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2476344Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2476574Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2476612Z graph_break [] 2025-12-04T09:49:26.2476686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2476726Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2476831Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2476903Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2477129Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2477166Z graph_break [] 2025-12-04T09:49:26.2477239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2477280Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2477334Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2477439Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2477665Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2477705Z graph_break [] 2025-12-04T09:49:26.2477778Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2477819Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2477873Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2477945Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2478172Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2478209Z graph_break [] 2025-12-04T09:49:26.2478283Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2478326Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2478381Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2478455Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2478683Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2478721Z graph_break [] 2025-12-04T09:49:26.2478806Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2478852Z Traceback (most recent call last): 2025-12-04T09:49:26.2478971Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2479010Z return value(self) 2025-12-04T09:49:26.2479155Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2479207Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2479359Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2479435Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2479593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2479667Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2479719Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2479721Z 2025-12-04T09:49:26.2479773Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2479889Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2479986Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2479989Z 2025-12-04T09:49:26.2480063Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2480214Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2480217Z 2025-12-04T09:49:26.2480305Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2480379Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2480422Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2480478Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2480708Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2480781Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2480818Z graph_break [] 2025-12-04T09:49:26.2480915Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2480959Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2481013Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2481086Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2481314Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2481351Z graph_break [] 2025-12-04T09:49:26.2481423Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2481466Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2481520Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2481593Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2481822Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2481862Z graph_break [] 2025-12-04T09:49:26.2481934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2481977Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2482031Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2482103Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2482329Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2482367Z graph_break [] 2025-12-04T09:49:26.2482439Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2482483Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2482537Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2482622Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2482861Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2482899Z graph_break [] 2025-12-04T09:49:26.2482971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2483013Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2483067Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2483140Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2483367Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2483406Z graph_break [] 2025-12-04T09:49:26.2483478Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2483521Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2483575Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2483647Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2483875Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2483911Z graph_break [] 2025-12-04T09:49:26.2483984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2484025Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2484079Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2484172Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2484403Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2484441Z graph_break [] 2025-12-04T09:49:26.2484514Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2484554Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2484610Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2484680Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2484908Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2484944Z graph_break [] 2025-12-04T09:49:26.2485019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2485061Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2485116Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2485188Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2485415Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2485451Z graph_break [] 2025-12-04T09:49:26.2485525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2485565Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2485620Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2485691Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2485920Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2485967Z graph_break [] 2025-12-04T09:49:26.2486054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2486095Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2486151Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2486222Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2486452Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2486488Z graph_break [] 2025-12-04T09:49:26.2486562Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2486602Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2486659Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2486731Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2486992Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2487028Z graph_break [] 2025-12-04T09:49:26.2487102Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2487142Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2487197Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2487268Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2487530Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2487569Z graph_break [] 2025-12-04T09:49:26.2487642Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2487685Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2487738Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2487812Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2488040Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2488080Z graph_break [] 2025-12-04T09:49:26.2488153Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2488197Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2488252Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2488329Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2488557Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2488600Z graph_break [] 2025-12-04T09:49:26.2488673Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2488717Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2488774Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2488848Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2489075Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2489114Z graph_break [] 2025-12-04T09:49:26.2489190Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2489235Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2489305Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2489392Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2489619Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2489658Z graph_break [] 2025-12-04T09:49:26.2489732Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2489777Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2489831Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2489906Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2490135Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2490175Z graph_break [] 2025-12-04T09:49:26.2490250Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2490294Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2490349Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2490425Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2490656Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2490696Z graph_break [] 2025-12-04T09:49:26.2490769Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2490815Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2490889Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2490966Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2491195Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2491235Z graph_break [] 2025-12-04T09:49:26.2491310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2491352Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2491407Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2491482Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2491709Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2491748Z graph_break [] 2025-12-04T09:49:26.2491823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2491866Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2491924Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2491996Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2492229Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2492266Z graph_break [] 2025-12-04T09:49:26.2492354Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2492401Z Traceback (most recent call last): 2025-12-04T09:49:26.2492522Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2492562Z return value(self) 2025-12-04T09:49:26.2492709Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2492772Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2492922Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2492983Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2493143Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2493215Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2493273Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2493275Z 2025-12-04T09:49:26.2493325Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2493445Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2493545Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2493548Z 2025-12-04T09:49:26.2493627Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2493779Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2493782Z 2025-12-04T09:49:26.2493873Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2493945Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2493990Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2494047Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2494279Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2494375Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2494414Z graph_break [] 2025-12-04T09:49:26.2494489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2494533Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2494591Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2494663Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2494894Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2494931Z graph_break [] 2025-12-04T09:49:26.2495007Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2495049Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2495109Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2495182Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2495414Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2495452Z graph_break [] 2025-12-04T09:49:26.2495527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2495570Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2495626Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2495698Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2496067Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2496105Z graph_break [] 2025-12-04T09:49:26.2496185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2496244Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2496301Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2496384Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2496614Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2496651Z graph_break [] 2025-12-04T09:49:26.2496727Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2496836Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2496892Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2496964Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2497195Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2497235Z graph_break [] 2025-12-04T09:49:26.2497311Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2497354Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2497410Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2497486Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2497716Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2497754Z graph_break [] 2025-12-04T09:49:26.2497830Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2497870Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2497964Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2498038Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2498269Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2498308Z graph_break [] 2025-12-04T09:49:26.2498382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2498425Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2498481Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2498555Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2498784Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2498824Z graph_break [] 2025-12-04T09:49:26.2498898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2498945Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2498999Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2499074Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2499300Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2499341Z graph_break [] 2025-12-04T09:49:26.2499414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2499461Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2499516Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2499593Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2499821Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2499887Z graph_break [] 2025-12-04T09:49:26.2499962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2500009Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2500065Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2500139Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2500365Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2500404Z graph_break [] 2025-12-04T09:49:26.2500481Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2500525Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2500580Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2500658Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2500887Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2500926Z graph_break [] 2025-12-04T09:49:26.2501000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2501044Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2501100Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2501175Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2501427Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2501469Z graph_break [] 2025-12-04T09:49:26.2501543Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2501587Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2501642Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2501717Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2501943Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2501982Z graph_break [] 2025-12-04T09:49:26.2502056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2502098Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2502153Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2502227Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2502459Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2502497Z graph_break [] 2025-12-04T09:49:26.2502573Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2502614Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2502672Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2502744Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2502973Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2503012Z graph_break [] 2025-12-04T09:49:26.2503097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2503140Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2503212Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2503287Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2503516Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2503553Z graph_break [] 2025-12-04T09:49:26.2503628Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2503670Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2503727Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2503800Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2504033Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2504074Z graph_break [] 2025-12-04T09:49:26.2504149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2504191Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2504249Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2504321Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2504554Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2504590Z graph_break [] 2025-12-04T09:49:26.2504687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2504729Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2504786Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2504859Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2505089Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2505126Z graph_break [] 2025-12-04T09:49:26.2505203Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2505244Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2505301Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2505373Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2505605Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2505644Z graph_break [] 2025-12-04T09:49:26.2505719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2505761Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2505818Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2505890Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2506123Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2506161Z graph_break [] 2025-12-04T09:49:26.2506234Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2506276Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2506334Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2506409Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2506650Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2506700Z graph_break [] 2025-12-04T09:49:26.2506825Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2506873Z Traceback (most recent call last): 2025-12-04T09:49:26.2506992Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2507034Z return value(self) 2025-12-04T09:49:26.2507179Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2507232Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2507371Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2507435Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2507595Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2507671Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2507726Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2507729Z 2025-12-04T09:49:26.2507783Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2507900Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2508002Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2508004Z 2025-12-04T09:49:26.2508077Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2508274Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2508278Z 2025-12-04T09:49:26.2508365Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2508443Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2508486Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2508544Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2508775Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2508853Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2508892Z graph_break [] 2025-12-04T09:49:26.2508964Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2509013Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2509067Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2509144Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2509374Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2509412Z graph_break [] 2025-12-04T09:49:26.2509485Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2509528Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2509583Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2509655Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2509884Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2509950Z graph_break [] 2025-12-04T09:49:26.2510023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2510078Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2510132Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2510208Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2510437Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2510475Z graph_break [] 2025-12-04T09:49:26.2510549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2510592Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2510647Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2510724Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2510954Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2510993Z graph_break [] 2025-12-04T09:49:26.2511066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2511110Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2511164Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2511239Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2511468Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2511508Z graph_break [] 2025-12-04T09:49:26.2511602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2511648Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2511703Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2511777Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2512004Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2512044Z graph_break [] 2025-12-04T09:49:26.2512117Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2512160Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2512214Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2512287Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2512516Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2512555Z graph_break [] 2025-12-04T09:49:26.2512630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2512670Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2512726Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2512797Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2513026Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2513062Z graph_break [] 2025-12-04T09:49:26.2513138Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2513179Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2513235Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2513316Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2513561Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2513597Z graph_break [] 2025-12-04T09:49:26.2513671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2513713Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2513768Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2513840Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2514071Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2514108Z graph_break [] 2025-12-04T09:49:26.2514184Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2514227Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2514283Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2514354Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2514584Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2514621Z graph_break [] 2025-12-04T09:49:26.2514693Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2514734Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2514789Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2514880Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2515108Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2515147Z graph_break [] 2025-12-04T09:49:26.2515220Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2515260Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2515316Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2515388Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2515615Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2515652Z graph_break [] 2025-12-04T09:49:26.2515728Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2515769Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2515825Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2515898Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2516125Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2516161Z graph_break [] 2025-12-04T09:49:26.2516235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2516276Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2516331Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2516401Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2516631Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2516679Z graph_break [] 2025-12-04T09:49:26.2516808Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2516850Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2516904Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2516976Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2517202Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2517240Z graph_break [] 2025-12-04T09:49:26.2517312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2517355Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2517410Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2517483Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2517710Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2517748Z graph_break [] 2025-12-04T09:49:26.2517820Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2517863Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2517916Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2517988Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2518218Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2518286Z graph_break [] 2025-12-04T09:49:26.2518361Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2518404Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2518459Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2518532Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2518757Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2518795Z graph_break [] 2025-12-04T09:49:26.2518867Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2518910Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2518963Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2519035Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2519265Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2519305Z graph_break [] 2025-12-04T09:49:26.2519378Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2519420Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2519474Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2519547Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2519774Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2519811Z graph_break [] 2025-12-04T09:49:26.2519883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2519927Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2519996Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2520068Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2520310Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2520346Z graph_break [] 2025-12-04T09:49:26.2520419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2520460Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2520515Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2520586Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2520815Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2520853Z graph_break [] 2025-12-04T09:49:26.2520925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2520968Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2521022Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2521093Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2521323Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2521358Z graph_break [] 2025-12-04T09:49:26.2521445Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2521489Z Traceback (most recent call last): 2025-12-04T09:49:26.2521630Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2521670Z return value(self) 2025-12-04T09:49:26.2521815Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2521866Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2522003Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2522064Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2522224Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2522296Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2522350Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2522352Z 2025-12-04T09:49:26.2522403Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2522522Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2522620Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2522624Z 2025-12-04T09:49:26.2522697Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2522847Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2522851Z 2025-12-04T09:49:26.2522938Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2523011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2523053Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2523110Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2523340Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2523427Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2523473Z graph_break [] 2025-12-04T09:49:26.2523547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2523588Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2523644Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2523715Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2523945Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2523981Z graph_break [] 2025-12-04T09:49:26.2524054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2524094Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2524150Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2524222Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2524453Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2524489Z graph_break [] 2025-12-04T09:49:26.2524563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2524603Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2524658Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2524729Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2524977Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2525014Z graph_break [] 2025-12-04T09:49:26.2525089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2525132Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2525188Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2525258Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2525486Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2525522Z graph_break [] 2025-12-04T09:49:26.2525595Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2525635Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2525690Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2525763Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2525991Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2526030Z graph_break [] 2025-12-04T09:49:26.2526105Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2526145Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2526201Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2526272Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2526500Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2526536Z graph_break [] 2025-12-04T09:49:26.2526611Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2526667Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2526722Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2526839Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2527067Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2527105Z graph_break [] 2025-12-04T09:49:26.2527177Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2527219Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2527273Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2527345Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2527574Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2527613Z graph_break [] 2025-12-04T09:49:26.2527686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2527727Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2527781Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2527854Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2528082Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2528119Z graph_break [] 2025-12-04T09:49:26.2528191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2528232Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2528319Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2528393Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2528620Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2528658Z graph_break [] 2025-12-04T09:49:26.2528730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2528772Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2528826Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2528898Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2529124Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2529163Z graph_break [] 2025-12-04T09:49:26.2529236Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2529279Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2529333Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2529406Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2529632Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2529670Z graph_break [] 2025-12-04T09:49:26.2529742Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2529784Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2529837Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2529910Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2530139Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2530202Z graph_break [] 2025-12-04T09:49:26.2530275Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2530317Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2530371Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2530443Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2530671Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2530706Z graph_break [] 2025-12-04T09:49:26.2530780Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2530822Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2530878Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2530949Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2531179Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2531215Z graph_break [] 2025-12-04T09:49:26.2531287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2531328Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2531382Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2531453Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2531702Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2531739Z graph_break [] 2025-12-04T09:49:26.2531813Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2531855Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2531909Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2531981Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2532211Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2532247Z graph_break [] 2025-12-04T09:49:26.2532321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2532361Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2532420Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2532494Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2532726Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2532764Z graph_break [] 2025-12-04T09:49:26.2541672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2541734Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2541797Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2541876Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2542121Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2542160Z graph_break [] 2025-12-04T09:49:26.2542245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2542334Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2542408Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2542483Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2542716Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2542755Z graph_break [] 2025-12-04T09:49:26.2542834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2542877Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2542937Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2543010Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2543244Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2543286Z graph_break [] 2025-12-04T09:49:26.2543362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2543405Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2543462Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2543535Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2543770Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2543807Z graph_break [] 2025-12-04T09:49:26.2543886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2543958Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2544020Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2544093Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2544327Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2544364Z graph_break [] 2025-12-04T09:49:26.2544440Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2544483Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2544540Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2544613Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2544848Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2544889Z graph_break [] 2025-12-04T09:49:26.2544963Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2545007Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2545062Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2545135Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2545364Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2545401Z graph_break [] 2025-12-04T09:49:26.2545488Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2545538Z Traceback (most recent call last): 2025-12-04T09:49:26.2545670Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2545713Z return value(self) 2025-12-04T09:49:26.2545877Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2545943Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2546088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2546152Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2546315Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2546391Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2546445Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2546448Z 2025-12-04T09:49:26.2546503Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2546624Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2546724Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2546727Z 2025-12-04T09:49:26.2546857Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2547014Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2547017Z 2025-12-04T09:49:26.2547106Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2547183Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2547226Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2547284Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2547547Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2547626Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2547663Z graph_break [] 2025-12-04T09:49:26.2547740Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2547782Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2547838Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2547910Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2548142Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2548179Z graph_break [] 2025-12-04T09:49:26.2548255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2548296Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2548354Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2548433Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2548665Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2548704Z graph_break [] 2025-12-04T09:49:26.2548777Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2548821Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2548878Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2548952Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2549183Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2549222Z graph_break [] 2025-12-04T09:49:26.2549312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2549374Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2549428Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2549501Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2549729Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2549766Z graph_break [] 2025-12-04T09:49:26.2549847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2549889Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2549945Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2550020Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2550250Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2550290Z graph_break [] 2025-12-04T09:49:26.2550363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2550405Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2550460Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2550535Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2550765Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2550803Z graph_break [] 2025-12-04T09:49:26.2550895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2550940Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2550996Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2551071Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2551298Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2551336Z graph_break [] 2025-12-04T09:49:26.2551409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2551451Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2551506Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2551580Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2551811Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2551850Z graph_break [] 2025-12-04T09:49:26.2551926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2551969Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2552024Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2552095Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2552325Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2552362Z graph_break [] 2025-12-04T09:49:26.2552435Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2552477Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2552534Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2552606Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2552857Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2552904Z graph_break [] 2025-12-04T09:49:26.2552980Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2553023Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2553078Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2553150Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2553380Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2553420Z graph_break [] 2025-12-04T09:49:26.2553495Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2553537Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2553595Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2553667Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2553899Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2553935Z graph_break [] 2025-12-04T09:49:26.2554009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2554049Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2554106Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2554178Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2554428Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2554468Z graph_break [] 2025-12-04T09:49:26.2554544Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2554584Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2554641Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2554713Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2554943Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2554979Z graph_break [] 2025-12-04T09:49:26.2555053Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2555097Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2555154Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2555227Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2555458Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2555497Z graph_break [] 2025-12-04T09:49:26.2555570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2555612Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2555667Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2555739Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2555969Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2556019Z graph_break [] 2025-12-04T09:49:26.2556093Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2556146Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2556201Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2556274Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2556504Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2556542Z graph_break [] 2025-12-04T09:49:26.2556615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2556658Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2556712Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2556898Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2557128Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2557167Z graph_break [] 2025-12-04T09:49:26.2557241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2557285Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2557340Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2557416Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2557647Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2557687Z graph_break [] 2025-12-04T09:49:26.2557791Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2557837Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2557894Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2557967Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2558195Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2558233Z graph_break [] 2025-12-04T09:49:26.2558307Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2558352Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2558407Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2558480Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2558714Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2558754Z graph_break [] 2025-12-04T09:49:26.2558827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2558869Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2558923Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2558997Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2559226Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2559266Z graph_break [] 2025-12-04T09:49:26.2559339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2559383Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2559437Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2559532Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2559776Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2559813Z graph_break [] 2025-12-04T09:49:26.2559889Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2559930Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2559986Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2560058Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2560289Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2560327Z graph_break [] 2025-12-04T09:49:26.2560402Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2560444Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2560501Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2560573Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2560806Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2560842Z graph_break [] 2025-12-04T09:49:26.2560918Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2560958Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2561014Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2561113Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2561345Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2561384Z graph_break [] 2025-12-04T09:49:26.2561472Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2561518Z Traceback (most recent call last): 2025-12-04T09:49:26.2561642Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2561682Z return value(self) 2025-12-04T09:49:26.2561829Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2561881Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2562026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2562087Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2562253Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2562329Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2562385Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2562387Z 2025-12-04T09:49:26.2562440Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2562559Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2562658Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2562661Z 2025-12-04T09:49:26.2562736Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2562893Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2562908Z 2025-12-04T09:49:26.2562996Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2563082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2563124Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2563180Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2563411Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2563484Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2563520Z graph_break [] 2025-12-04T09:49:26.2563595Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2563636Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2563694Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2563770Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2564003Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2564042Z graph_break [] 2025-12-04T09:49:26.2564123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2564165Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2564223Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2564297Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2564530Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2564610Z graph_break [] 2025-12-04T09:49:26.2564687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2564730Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2564788Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2564862Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2565093Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2565131Z graph_break [] 2025-12-04T09:49:26.2565209Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2565251Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2565308Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2565381Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2565612Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2565651Z graph_break [] 2025-12-04T09:49:26.2565727Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2565768Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2565826Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2565899Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2566128Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2566169Z graph_break [] 2025-12-04T09:49:26.2566243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2566288Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2566359Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2566433Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2566673Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2566711Z graph_break [] 2025-12-04T09:49:26.2566845Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2566888Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2566943Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2567018Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2567247Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2567288Z graph_break [] 2025-12-04T09:49:26.2567361Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2567403Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2567462Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2567534Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2567764Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2567800Z graph_break [] 2025-12-04T09:49:26.2567875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2567917Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2567974Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2568092Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2568327Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2568364Z graph_break [] 2025-12-04T09:49:26.2568440Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2568481Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2568538Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2568610Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2568841Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2568878Z graph_break [] 2025-12-04T09:49:26.2568957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2568999Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2569058Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2569129Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2569359Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2569397Z graph_break [] 2025-12-04T09:49:26.2569473Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2569514Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2569572Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2569646Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2569881Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2569947Z graph_break [] 2025-12-04T09:49:26.2570023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2570064Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2570124Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2570198Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2570430Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2570468Z graph_break [] 2025-12-04T09:49:26.2570542Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2570586Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2570644Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2570717Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2570950Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2570990Z graph_break [] 2025-12-04T09:49:26.2571063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2571108Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2571163Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2571237Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2571489Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2571531Z graph_break [] 2025-12-04T09:49:26.2571606Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2571652Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2571711Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2571783Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2572016Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2572054Z graph_break [] 2025-12-04T09:49:26.2572129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2572169Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2572226Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2572300Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2572533Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2572571Z graph_break [] 2025-12-04T09:49:26.2572647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2572689Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2572748Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2572818Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2573052Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2573090Z graph_break [] 2025-12-04T09:49:26.2573167Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2573223Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2573281Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2573364Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2573595Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2573632Z graph_break [] 2025-12-04T09:49:26.2573705Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2573749Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2573805Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2573878Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2574107Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2574149Z graph_break [] 2025-12-04T09:49:26.2574222Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2574269Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2574324Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2574398Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2574626Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2574665Z graph_break [] 2025-12-04T09:49:26.2574738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2574782Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2574859Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2574934Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2575169Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2575208Z graph_break [] 2025-12-04T09:49:26.2575282Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2575326Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2575381Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2575455Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2575685Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2575724Z graph_break [] 2025-12-04T09:49:26.2575798Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2575843Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2575899Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2575974Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2576202Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2576242Z graph_break [] 2025-12-04T09:49:26.2576315Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2576360Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2576414Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2576490Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2576731Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2576828Z graph_break [] 2025-12-04T09:49:26.2576902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2576947Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2577003Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2577077Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2577311Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2577352Z graph_break [] 2025-12-04T09:49:26.2577429Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2577471Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2577526Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2577602Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2577829Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2577867Z graph_break [] 2025-12-04T09:49:26.2577956Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2578003Z Traceback (most recent call last): 2025-12-04T09:49:26.2578126Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2578165Z return value(self) 2025-12-04T09:49:26.2578350Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2578403Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2578543Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2578604Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2578766Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2578838Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2578892Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2578894Z 2025-12-04T09:49:26.2578946Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2579065Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2579163Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2579167Z 2025-12-04T09:49:26.2579244Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2579399Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2579402Z 2025-12-04T09:49:26.2579491Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2579566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2579612Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2579667Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2579898Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2579971Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2580011Z graph_break [] 2025-12-04T09:49:26.2580085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2580145Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2580216Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2580290Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2580521Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2580561Z graph_break [] 2025-12-04T09:49:26.2580635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2580677Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2580734Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2580806Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2581037Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2581075Z graph_break [] 2025-12-04T09:49:26.2581149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2581190Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2581246Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2581318Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2581548Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2581585Z graph_break [] 2025-12-04T09:49:26.2581659Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2581721Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2581778Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2581849Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2582078Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2582114Z graph_break [] 2025-12-04T09:49:26.2582188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2582228Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2582283Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2582355Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2582585Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2582622Z graph_break [] 2025-12-04T09:49:26.2582695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2582736Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2582791Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2582861Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2583090Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2583126Z graph_break [] 2025-12-04T09:49:26.2583202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2583242Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2583298Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2583370Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2583610Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2583657Z graph_break [] 2025-12-04T09:49:26.2583731Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2583771Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2583826Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2583897Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2584125Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2584163Z graph_break [] 2025-12-04T09:49:26.2584238Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2584281Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2584335Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2584409Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2584636Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2584674Z graph_break [] 2025-12-04T09:49:26.2584747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2584789Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2584845Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2584917Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2585169Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2585210Z graph_break [] 2025-12-04T09:49:26.2585283Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2585325Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2585380Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2585453Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2585681Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2585718Z graph_break [] 2025-12-04T09:49:26.2585791Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2585834Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2585890Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2585964Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2586193Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2586230Z graph_break [] 2025-12-04T09:49:26.2586303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2586344Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2586398Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2586471Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2586699Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2586805Z graph_break [] 2025-12-04T09:49:26.2586878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2586936Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2586990Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2587063Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2587292Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2587329Z graph_break [] 2025-12-04T09:49:26.2587401Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2587442Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2587496Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2587570Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2587799Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2587837Z graph_break [] 2025-12-04T09:49:26.2587911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2587952Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2588006Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2588077Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2588305Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2588341Z graph_break [] 2025-12-04T09:49:26.2588444Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2588486Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2588541Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2588613Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2588840Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2588877Z graph_break [] 2025-12-04T09:49:26.2588949Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2588989Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2589044Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2589114Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2589348Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2589385Z graph_break [] 2025-12-04T09:49:26.2589460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2589501Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2589556Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2589626Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2589854Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2589890Z graph_break [] 2025-12-04T09:49:26.2589962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2590003Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2590059Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2590145Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2590388Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2590424Z graph_break [] 2025-12-04T09:49:26.2590498Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2590538Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2590593Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2590664Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2590895Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2590931Z graph_break [] 2025-12-04T09:49:26.2591005Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2591047Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2591102Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2591173Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2591404Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2591440Z graph_break [] 2025-12-04T09:49:26.2591513Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2591554Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2591609Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2591698Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2591928Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2591966Z graph_break [] 2025-12-04T09:49:26.2592038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2592080Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2592135Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2592207Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2592434Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2592470Z graph_break [] 2025-12-04T09:49:26.2592544Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2592585Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2592640Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2592713Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2592940Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2592977Z graph_break [] 2025-12-04T09:49:26.2593049Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2593091Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2593145Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2593218Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2593447Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2593497Z graph_break [] 2025-12-04T09:49:26.2593580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2593622Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2593677Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2593749Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2593976Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2594014Z graph_break [] 2025-12-04T09:49:26.2594086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2594128Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2594183Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2594256Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2594485Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2594524Z graph_break [] 2025-12-04T09:49:26.2594609Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2594657Z Traceback (most recent call last): 2025-12-04T09:49:26.2594777Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2594817Z return value(self) 2025-12-04T09:49:26.2594963Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2595016Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2595173Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2595237Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2595399Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2595471Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2595526Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2595528Z 2025-12-04T09:49:26.2595579Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2595697Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2595794Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2595797Z 2025-12-04T09:49:26.2595871Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2596024Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2596028Z 2025-12-04T09:49:26.2596115Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2596189Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2596231Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2596286Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2596516Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2596588Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2596626Z graph_break [] 2025-12-04T09:49:26.2596698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2596787Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2596843Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2596931Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2597175Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2597212Z graph_break [] 2025-12-04T09:49:26.2597285Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2597328Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2597381Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2597453Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2597684Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2597723Z graph_break [] 2025-12-04T09:49:26.2597796Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2597840Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2597894Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2597966Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2598194Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2598231Z graph_break [] 2025-12-04T09:49:26.2598304Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2598345Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2598399Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2598508Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2598737Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2598774Z graph_break [] 2025-12-04T09:49:26.2598848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2598888Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2598943Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2599013Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2599241Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2599277Z graph_break [] 2025-12-04T09:49:26.2599353Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2599395Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2599450Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2599522Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2599751Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2599787Z graph_break [] 2025-12-04T09:49:26.2599861Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2599901Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2599955Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2600026Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2600257Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2600319Z graph_break [] 2025-12-04T09:49:26.2600392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2600432Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2600486Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2600557Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2600785Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2600820Z graph_break [] 2025-12-04T09:49:26.2600894Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2600934Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2600990Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2601062Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2601292Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2601328Z graph_break [] 2025-12-04T09:49:26.2601403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2601444Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2601500Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2601572Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2601823Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2601860Z graph_break [] 2025-12-04T09:49:26.2601935Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2601977Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2602033Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2602104Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2602334Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2602372Z graph_break [] 2025-12-04T09:49:26.2602445Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2602487Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2602542Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2602616Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2602844Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2602884Z graph_break [] 2025-12-04T09:49:26.2602957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2602999Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2603054Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2603126Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2603357Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2603395Z graph_break [] 2025-12-04T09:49:26.2603470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2603531Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2603586Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2603683Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2603912Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2603950Z graph_break [] 2025-12-04T09:49:26.2604022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2604064Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2604118Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2604191Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2604420Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2604458Z graph_break [] 2025-12-04T09:49:26.2604531Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2604573Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2604626Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2604698Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2604925Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2604961Z graph_break [] 2025-12-04T09:49:26.2605033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2605074Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2605148Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2605221Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2605451Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2605490Z graph_break [] 2025-12-04T09:49:26.2605562Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2605604Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2605658Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2605730Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2605958Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2605996Z graph_break [] 2025-12-04T09:49:26.2606071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2606111Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2606167Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2606239Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2606468Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2606504Z graph_break [] 2025-12-04T09:49:26.2606578Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2606618Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2606673Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2606769Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2607001Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2607076Z graph_break [] 2025-12-04T09:49:26.2607152Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2607192Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2607248Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2607319Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2607551Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2607586Z graph_break [] 2025-12-04T09:49:26.2607661Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2607702Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2607759Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2607830Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2608060Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2608096Z graph_break [] 2025-12-04T09:49:26.2608171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2608212Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2608267Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2608338Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2608592Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2608630Z graph_break [] 2025-12-04T09:49:26.2608704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2608745Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2608801Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2608872Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2609101Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2609137Z graph_break [] 2025-12-04T09:49:26.2609212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2609253Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2609311Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2609383Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2609615Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2609653Z graph_break [] 2025-12-04T09:49:26.2609726Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2609767Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2609822Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2609894Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2610121Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2610160Z graph_break [] 2025-12-04T09:49:26.2610234Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2610291Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2610357Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2610429Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2610657Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2610694Z graph_break [] 2025-12-04T09:49:26.2610767Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2610809Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2610863Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2610936Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2611166Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2611205Z graph_break [] 2025-12-04T09:49:26.2611279Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2611320Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2611375Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2611449Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2611678Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2611715Z graph_break [] 2025-12-04T09:49:26.2611800Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2611874Z Traceback (most recent call last): 2025-12-04T09:49:26.2611996Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2612038Z return value(self) 2025-12-04T09:49:26.2612182Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2612234Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2612372Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2612434Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2612594Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2612668Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2612721Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2612723Z 2025-12-04T09:49:26.2612778Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2612895Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2612994Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2612996Z 2025-12-04T09:49:26.2613070Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2613222Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2613224Z 2025-12-04T09:49:26.2613312Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2613386Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2613428Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2613483Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2613718Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2613813Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2613850Z graph_break [] 2025-12-04T09:49:26.2613923Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2613965Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2614020Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2614092Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2614321Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2614358Z graph_break [] 2025-12-04T09:49:26.2614432Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2614474Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2614528Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2614602Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2614829Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2614866Z graph_break [] 2025-12-04T09:49:26.2614939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2614981Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2615036Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2615108Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2615366Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2615404Z graph_break [] 2025-12-04T09:49:26.2615479Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2615520Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2615575Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2615647Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2615875Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2615912Z graph_break [] 2025-12-04T09:49:26.2615984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2616029Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2616085Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2616161Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2616393Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2616430Z graph_break [] 2025-12-04T09:49:26.2616506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2616546Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2616602Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2616675Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2616933Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2616970Z graph_break [] 2025-12-04T09:49:26.2617064Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2617118Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2617177Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2617249Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2617479Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2617517Z graph_break [] 2025-12-04T09:49:26.2617591Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2617632Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2617689Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2617763Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2617993Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2618032Z graph_break [] 2025-12-04T09:49:26.2618106Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2618147Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2618205Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2618278Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2618508Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2618545Z graph_break [] 2025-12-04T09:49:26.2618645Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2618687Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2618746Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2618822Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2619051Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2619089Z graph_break [] 2025-12-04T09:49:26.2619165Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2619208Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2619266Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2619339Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2619572Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2619610Z graph_break [] 2025-12-04T09:49:26.2619689Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2619733Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2619790Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2619861Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2620093Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2620132Z graph_break [] 2025-12-04T09:49:26.2620206Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2620250Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2620308Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2620382Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2620621Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2620670Z graph_break [] 2025-12-04T09:49:26.2620742Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2620786Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2620842Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2620916Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2621146Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2621186Z graph_break [] 2025-12-04T09:49:26.2621260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2621306Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2621362Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2621436Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2621668Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2621709Z graph_break [] 2025-12-04T09:49:26.2621784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2621828Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2621883Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2621957Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2622206Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2622248Z graph_break [] 2025-12-04T09:49:26.2622321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2622364Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2622420Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2622493Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2622722Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2622762Z graph_break [] 2025-12-04T09:49:26.2622837Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2622881Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2622937Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2623013Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2623243Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2623283Z graph_break [] 2025-12-04T09:49:26.2623357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2623400Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2623456Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2623530Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2623762Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2623812Z graph_break [] 2025-12-04T09:49:26.2623888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2623940Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2623998Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2624069Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2624299Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2624335Z graph_break [] 2025-12-04T09:49:26.2624410Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2624451Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2624508Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2624582Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2624816Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2624854Z graph_break [] 2025-12-04T09:49:26.2624929Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2624970Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2625026Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2625098Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2625330Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2625365Z graph_break [] 2025-12-04T09:49:26.2625463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2625506Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2625567Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2625639Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2625874Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2625911Z graph_break [] 2025-12-04T09:49:26.2625988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2626029Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2626086Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2626159Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2626391Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2626430Z graph_break [] 2025-12-04T09:49:26.2626508Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2626551Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2626608Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2626681Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2626942Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2626979Z graph_break [] 2025-12-04T09:49:26.2627055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2627097Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2627155Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2627242Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2627485Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2627526Z graph_break [] 2025-12-04T09:49:26.2627600Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2627642Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2627698Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2627773Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2628006Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2628048Z graph_break [] 2025-12-04T09:49:26.2628121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2628166Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2628221Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2628295Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2628522Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2628562Z graph_break [] 2025-12-04T09:49:26.2628635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2628679Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2628734Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2628841Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2629073Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2629113Z graph_break [] 2025-12-04T09:49:26.2629187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2629231Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2629287Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2629361Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2629588Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2629629Z graph_break [] 2025-12-04T09:49:26.2629717Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2629767Z Traceback (most recent call last): 2025-12-04T09:49:26.2629887Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2629931Z return value(self) 2025-12-04T09:49:26.2630077Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2630133Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2630272Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2630336Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2630495Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2630570Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2630626Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2630638Z 2025-12-04T09:49:26.2630692Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2630821Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2630922Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2630924Z 2025-12-04T09:49:26.2630999Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2631148Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2631151Z 2025-12-04T09:49:26.2631238Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2631312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2631355Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2631411Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2631642Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2631715Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2631752Z graph_break [] 2025-12-04T09:49:26.2631826Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2631869Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2631925Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2631997Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2632226Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2632281Z graph_break [] 2025-12-04T09:49:26.2632355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2632398Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2632454Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2632526Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2632754Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2632791Z graph_break [] 2025-12-04T09:49:26.2632863Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2632904Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2632958Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2633030Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2633258Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2633298Z graph_break [] 2025-12-04T09:49:26.2633370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2633412Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2633465Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2633539Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2633766Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2633803Z graph_break [] 2025-12-04T09:49:26.2633876Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2633920Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2633984Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2634057Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2634295Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2634332Z graph_break [] 2025-12-04T09:49:26.2634406Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2634446Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2634501Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2634572Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2634801Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2634838Z graph_break [] 2025-12-04T09:49:26.2634912Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2634953Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2635009Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2635080Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2635307Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2635343Z graph_break [] 2025-12-04T09:49:26.2635416Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2635456Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2635530Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2635602Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2635836Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2635874Z graph_break [] 2025-12-04T09:49:26.2635947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2635987Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2636042Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2636113Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2636341Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2636377Z graph_break [] 2025-12-04T09:49:26.2636454Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2636495Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2636553Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2636624Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2636905Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2636942Z graph_break [] 2025-12-04T09:49:26.2637015Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2637055Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2637111Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2637182Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2637411Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2637477Z graph_break [] 2025-12-04T09:49:26.2637551Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2637592Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2637649Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2637722Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2637951Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2637989Z graph_break [] 2025-12-04T09:49:26.2638062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2638105Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2638159Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2638231Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2638459Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2638495Z graph_break [] 2025-12-04T09:49:26.2638568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2638609Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2638664Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2638738Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2638992Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2639031Z graph_break [] 2025-12-04T09:49:26.2639104Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2639148Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2639202Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2639274Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2639501Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2639538Z graph_break [] 2025-12-04T09:49:26.2639610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2639653Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2639707Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2639784Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2640015Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2640054Z graph_break [] 2025-12-04T09:49:26.2640126Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2640167Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2640221Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2640292Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2640519Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2640556Z graph_break [] 2025-12-04T09:49:26.2640629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2640684Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2640739Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2640828Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2641055Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2641093Z graph_break [] 2025-12-04T09:49:26.2641165Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2641207Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2641261Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2641333Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2641562Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2641601Z graph_break [] 2025-12-04T09:49:26.2641674Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2641716Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2641770Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2641843Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2642072Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2642108Z graph_break [] 2025-12-04T09:49:26.2642182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2642223Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2642299Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2642372Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2642601Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2642638Z graph_break [] 2025-12-04T09:49:26.2642711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2642752Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2642808Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2642878Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2643108Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2643144Z graph_break [] 2025-12-04T09:49:26.2643218Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2643259Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2643315Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2643385Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2643612Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2643648Z graph_break [] 2025-12-04T09:49:26.2643722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2643763Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2643817Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2643889Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2644130Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2644177Z graph_break [] 2025-12-04T09:49:26.2644251Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2644291Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2644347Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2644418Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2644646Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2644683Z graph_break [] 2025-12-04T09:49:26.2644759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2644800Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2644856Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2644928Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2645156Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2645193Z graph_break [] 2025-12-04T09:49:26.2645266Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2645306Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2645361Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2645432Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2645680Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2645720Z graph_break [] 2025-12-04T09:49:26.2645795Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2645837Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2645892Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2645964Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2646191Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2646228Z graph_break [] 2025-12-04T09:49:26.2646300Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2646341Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2646396Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2646470Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2646696Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2646735Z graph_break [] 2025-12-04T09:49:26.2646862Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2646906Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2646961Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2647034Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2647261Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2647300Z graph_break [] 2025-12-04T09:49:26.2647402Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2647458Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2647513Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2647585Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2647814Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2647852Z graph_break [] 2025-12-04T09:49:26.2647936Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2647982Z Traceback (most recent call last): 2025-12-04T09:49:26.2648101Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2648143Z return value(self) 2025-12-04T09:49:26.2648286Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2648340Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2648479Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2648541Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2648700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2648773Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2648826Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2648829Z 2025-12-04T09:49:26.2648881Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2648998Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2649121Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2649124Z 2025-12-04T09:49:26.2649198Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2649349Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2649351Z 2025-12-04T09:49:26.2649438Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2649510Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2649553Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2649608Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2649838Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2649913Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2649950Z graph_break [] 2025-12-04T09:49:26.2650023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2650066Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2650121Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2650193Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2650423Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2650460Z graph_break [] 2025-12-04T09:49:26.2650533Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2650574Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2650629Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2650703Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2650944Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2650992Z graph_break [] 2025-12-04T09:49:26.2651065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2651107Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2651161Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2651234Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2651461Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2651499Z graph_break [] 2025-12-04T09:49:26.2651574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2651616Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2651672Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2651745Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2651975Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2652012Z graph_break [] 2025-12-04T09:49:26.2652086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2652128Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2652181Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2652253Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2652503Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2652541Z graph_break [] 2025-12-04T09:49:26.2652615Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2652655Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2652712Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2652783Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2653014Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2653050Z graph_break [] 2025-12-04T09:49:26.2653123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2653166Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2653221Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2653293Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2653523Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2653559Z graph_break [] 2025-12-04T09:49:26.2653633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2653674Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2653730Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2653801Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2654035Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2654092Z graph_break [] 2025-12-04T09:49:26.2654167Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2654217Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2654273Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2654344Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2654573Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2654609Z graph_break [] 2025-12-04T09:49:26.2654685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2654725Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2654780Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2654852Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2655081Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2655117Z graph_break [] 2025-12-04T09:49:26.2655191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2655231Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2655287Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2655358Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2655588Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2655624Z graph_break [] 2025-12-04T09:49:26.2655723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2655764Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2655819Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2655891Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2656120Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2656158Z graph_break [] 2025-12-04T09:49:26.2656230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2656271Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2656325Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2656397Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2656625Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2656667Z graph_break [] 2025-12-04T09:49:26.2656783Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2656826Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2656881Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2656953Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2657180Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2657218Z graph_break [] 2025-12-04T09:49:26.2657290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2657331Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2657387Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2657485Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2657737Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2657774Z graph_break [] 2025-12-04T09:49:26.2657847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2657888Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2657943Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2658015Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2658245Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2658282Z graph_break [] 2025-12-04T09:49:26.2658356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2658399Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2658453Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2658524Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2658750Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2658786Z graph_break [] 2025-12-04T09:49:26.2658858Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2658900Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2658954Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2659058Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2659286Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2659325Z graph_break [] 2025-12-04T09:49:26.2659397Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2659439Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2659493Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2659566Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2659794Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2659831Z graph_break [] 2025-12-04T09:49:26.2659908Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2659950Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2660006Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2660080Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2660308Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2660345Z graph_break [] 2025-12-04T09:49:26.2662642Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2662689Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2662747Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2662823Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2663069Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2663130Z graph_break [] 2025-12-04T09:49:26.2663217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2663258Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2663314Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2663385Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2663616Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2663652Z graph_break [] 2025-12-04T09:49:26.2663727Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2663768Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2663824Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2663896Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2664128Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2664166Z graph_break [] 2025-12-04T09:49:26.2664241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2664281Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2664337Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2664408Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2664635Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2664698Z graph_break [] 2025-12-04T09:49:26.2664776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2664820Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2664876Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2664947Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2665179Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2665214Z graph_break [] 2025-12-04T09:49:26.2665289Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2665329Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2665385Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2665456Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2665687Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2665729Z graph_break [] 2025-12-04T09:49:26.2665803Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2665843Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2665898Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2665970Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2666199Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2666236Z graph_break [] 2025-12-04T09:49:26.2666308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2666352Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2666420Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2666492Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2666731Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2666808Z graph_break [] 2025-12-04T09:49:26.2666881Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2666923Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2666977Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2667049Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2667278Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2667316Z graph_break [] 2025-12-04T09:49:26.2667388Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2667431Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2667485Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2667557Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2667784Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2667821Z graph_break [] 2025-12-04T09:49:26.2667894Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2667935Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2667990Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2668110Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2668340Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2668378Z graph_break [] 2025-12-04T09:49:26.2668451Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2668492Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2668547Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2668619Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2668847Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2668884Z graph_break [] 2025-12-04T09:49:26.2668971Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2669020Z Traceback (most recent call last): 2025-12-04T09:49:26.2669145Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2669189Z return value(self) 2025-12-04T09:49:26.2669336Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2669390Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2669531Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2669594Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2669755Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2669832Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2669886Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2669914Z 2025-12-04T09:49:26.2669967Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2670101Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2670201Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2670203Z 2025-12-04T09:49:26.2670277Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2670430Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2670433Z 2025-12-04T09:49:26.2670521Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2670596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2670639Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2670698Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2670930Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2671003Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2671040Z graph_break [] 2025-12-04T09:49:26.2671113Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2671155Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2671210Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2671282Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2671532Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2671571Z graph_break [] 2025-12-04T09:49:26.2671646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2671691Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2671745Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2671820Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2672046Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2672083Z graph_break [] 2025-12-04T09:49:26.2672156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2672199Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2672253Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2672327Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2672576Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2672614Z graph_break [] 2025-12-04T09:49:26.2672687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2672728Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2672783Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2672855Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2673082Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2673117Z graph_break [] 2025-12-04T09:49:26.2673192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2673246Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2673302Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2673383Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2673611Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2673647Z graph_break [] 2025-12-04T09:49:26.2673722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2673763Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2673819Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2673890Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2674122Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2674159Z graph_break [] 2025-12-04T09:49:26.2674235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2674274Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2674329Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2674400Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2674630Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2674666Z graph_break [] 2025-12-04T09:49:26.2674740Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2674780Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2674853Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2674926Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2675153Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2675191Z graph_break [] 2025-12-04T09:49:26.2675264Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2675305Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2675361Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2675431Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2675660Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2675697Z graph_break [] 2025-12-04T09:49:26.2675772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2675813Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2675870Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2675941Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2676168Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2676204Z graph_break [] 2025-12-04T09:49:26.2676278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2676318Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2676373Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2676446Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2676675Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2676770Z graph_break [] 2025-12-04T09:49:26.2676843Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2676884Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2676939Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2677011Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2677237Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2677274Z graph_break [] 2025-12-04T09:49:26.2677348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2677390Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2677446Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2677519Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2677749Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2677788Z graph_break [] 2025-12-04T09:49:26.2677861Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2677903Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2677957Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2678032Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2678288Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2678329Z graph_break [] 2025-12-04T09:49:26.2678403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2678447Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2678502Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2678575Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2678804Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2678841Z graph_break [] 2025-12-04T09:49:26.2678914Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2678957Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2679015Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2679087Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2679316Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2679355Z graph_break [] 2025-12-04T09:49:26.2679428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2679469Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2679523Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2679595Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2679823Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2679864Z graph_break [] 2025-12-04T09:49:26.2679937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2679995Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2680063Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2680136Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2680363Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2680401Z graph_break [] 2025-12-04T09:49:26.2680474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2680517Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2680571Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2680644Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2680872Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2680911Z graph_break [] 2025-12-04T09:49:26.2680985Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2681025Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2681082Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2681153Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2681382Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2681418Z graph_break [] 2025-12-04T09:49:26.2681491Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2681558Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2681616Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2681687Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2681916Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2681953Z graph_break [] 2025-12-04T09:49:26.2682028Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2682068Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2682123Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2682194Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2682425Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2682463Z graph_break [] 2025-12-04T09:49:26.2682536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2682578Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2682633Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2682705Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2682933Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2682969Z graph_break [] 2025-12-04T09:49:26.2683041Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2683082Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2683138Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2683210Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2683451Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2683498Z graph_break [] 2025-12-04T09:49:26.2683571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2683612Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2683666Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2683737Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2683965Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2684001Z graph_break [] 2025-12-04T09:49:26.2684077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2684120Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2684176Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2684248Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2684480Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2684522Z graph_break [] 2025-12-04T09:49:26.2684595Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2684637Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2684692Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2684763Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2685012Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2685054Z graph_break [] 2025-12-04T09:49:26.2685127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2685168Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2685224Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2685298Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2685525Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2685564Z graph_break [] 2025-12-04T09:49:26.2685636Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2685680Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2685735Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2685810Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2686040Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2686078Z graph_break [] 2025-12-04T09:49:26.2686150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2686191Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2686245Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2686316Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2686546Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2686599Z graph_break [] 2025-12-04T09:49:26.2686672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2686726Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2686817Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2686890Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2687120Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2687159Z graph_break [] 2025-12-04T09:49:26.2687231Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2687273Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2687329Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2687403Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2687633Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2687671Z graph_break [] 2025-12-04T09:49:26.2687745Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2687788Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2687846Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2687917Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2688147Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2688184Z graph_break [] 2025-12-04T09:49:26.2688309Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2688357Z Traceback (most recent call last): 2025-12-04T09:49:26.2688482Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2688523Z return value(self) 2025-12-04T09:49:26.2688670Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2688723Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2688861Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2688920Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2689080Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2689156Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2689214Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2689217Z 2025-12-04T09:49:26.2689270Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2689387Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2689488Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2689490Z 2025-12-04T09:49:26.2689563Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2689716Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2689718Z 2025-12-04T09:49:26.2689803Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2689879Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2689921Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2689984Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2690213Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2690317Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2690353Z graph_break [] 2025-12-04T09:49:26.2690428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2690470Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2690529Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2690601Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2690834Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2690874Z graph_break [] 2025-12-04T09:49:26.2690949Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2690991Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2691049Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2691122Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2691351Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2691388Z graph_break [] 2025-12-04T09:49:26.2691463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2691505Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2691560Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2691631Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2691882Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2691923Z graph_break [] 2025-12-04T09:49:26.2691998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2692040Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2692096Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2692167Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2692395Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2692431Z graph_break [] 2025-12-04T09:49:26.2692505Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2692547Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2692601Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2692672Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2692900Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2692936Z graph_break [] 2025-12-04T09:49:26.2693009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2693049Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2693105Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2693177Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2693407Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2693455Z graph_break [] 2025-12-04T09:49:26.2693529Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2693587Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2693643Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2693714Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2693941Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2693978Z graph_break [] 2025-12-04T09:49:26.2694050Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2694091Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2694145Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2694218Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2694446Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2694484Z graph_break [] 2025-12-04T09:49:26.2694557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2694598Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2694651Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2694723Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2694950Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2694987Z graph_break [] 2025-12-04T09:49:26.2695080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2695123Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2695178Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2695249Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2695475Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2695511Z graph_break [] 2025-12-04T09:49:26.2695582Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2695623Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2695676Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2695749Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2695977Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2696016Z graph_break [] 2025-12-04T09:49:26.2696088Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2696129Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2696182Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2696253Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2696479Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2696516Z graph_break [] 2025-12-04T09:49:26.2696588Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2696630Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2696684Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2696809Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2697050Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2697086Z graph_break [] 2025-12-04T09:49:26.2697158Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2697199Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2697253Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2697325Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2697554Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2697592Z graph_break [] 2025-12-04T09:49:26.2697664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2697707Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2697761Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2697833Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2698061Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2698098Z graph_break [] 2025-12-04T09:49:26.2698171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2698211Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2698266Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2698367Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2698600Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2698636Z graph_break [] 2025-12-04T09:49:26.2698709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2698749Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2698804Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2698875Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2699103Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2699139Z graph_break [] 2025-12-04T09:49:26.2699214Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2699256Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2699310Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2699382Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2699612Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2699647Z graph_break [] 2025-12-04T09:49:26.2699720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2699760Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2699815Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2699886Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2700117Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2700179Z graph_break [] 2025-12-04T09:49:26.2700252Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2700291Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2700346Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2700419Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2700649Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2700685Z graph_break [] 2025-12-04T09:49:26.2700759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2700800Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2700857Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2700929Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2701158Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2701195Z graph_break [] 2025-12-04T09:49:26.2701267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2701307Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2701362Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2701433Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2701679Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2701717Z graph_break [] 2025-12-04T09:49:26.2701790Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2701832Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2701886Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2701957Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2702185Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2702221Z graph_break [] 2025-12-04T09:49:26.2702294Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2702335Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2702389Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2702462Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2702690Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2702729Z graph_break [] 2025-12-04T09:49:26.2702801Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2702843Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2702896Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2702969Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2703197Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2703234Z graph_break [] 2025-12-04T09:49:26.2703308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2703365Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2703418Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2703504Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2703731Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2703768Z graph_break [] 2025-12-04T09:49:26.2703840Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2703880Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2703934Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2704005Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2704234Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2704271Z graph_break [] 2025-12-04T09:49:26.2704346Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2704387Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2704441Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2704512Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2704738Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2704775Z graph_break [] 2025-12-04T09:49:26.2704847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2704888Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2704965Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2705039Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2705265Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2705303Z graph_break [] 2025-12-04T09:49:26.2705375Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2705416Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2705470Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2705542Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2705771Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2705809Z graph_break [] 2025-12-04T09:49:26.2705883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2705923Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2705978Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2706049Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2706277Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2706313Z graph_break [] 2025-12-04T09:49:26.2706386Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2706426Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2706481Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2706552Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2706833Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2706913Z graph_break [] 2025-12-04T09:49:26.2706987Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2707027Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2707082Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2707153Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2707382Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2707418Z graph_break [] 2025-12-04T09:49:26.2707493Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2707533Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2707588Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2707659Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2707887Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2707923Z graph_break [] 2025-12-04T09:49:26.2708009Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2708055Z Traceback (most recent call last): 2025-12-04T09:49:26.2708177Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2708216Z return value(self) 2025-12-04T09:49:26.2708360Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2708436Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2708577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2708638Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2708799Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2708871Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2708924Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2708926Z 2025-12-04T09:49:26.2708977Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2709093Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2709192Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2709194Z 2025-12-04T09:49:26.2709268Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2709420Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2709423Z 2025-12-04T09:49:26.2709509Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2709583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2709624Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2709680Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2709908Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2709980Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2710016Z graph_break [] 2025-12-04T09:49:26.2710090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2710143Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2710198Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2710279Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2710511Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2710547Z graph_break [] 2025-12-04T09:49:26.2710621Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2710661Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2710718Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2710789Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2711019Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2711057Z graph_break [] 2025-12-04T09:49:26.2711130Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2711170Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2711225Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2711296Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2711525Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2711560Z graph_break [] 2025-12-04T09:49:26.2711633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2711697Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2711753Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2711825Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2712055Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2712092Z graph_break [] 2025-12-04T09:49:26.2712164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2712205Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2712259Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2712330Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2712565Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2712602Z graph_break [] 2025-12-04T09:49:26.2712675Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2712717Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2712771Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2712843Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2713070Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2713106Z graph_break [] 2025-12-04T09:49:26.2713179Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2713220Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2713273Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2713347Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2713586Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2713640Z graph_break [] 2025-12-04T09:49:26.2713713Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2713754Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2713808Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2713879Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2714106Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2714142Z graph_break [] 2025-12-04T09:49:26.2714216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2714258Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2714311Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2714384Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2714613Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2714650Z graph_break [] 2025-12-04T09:49:26.2714722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2714764Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2714817Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2714889Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2715138Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2715176Z graph_break [] 2025-12-04T09:49:26.2715249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2715291Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2715344Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2715417Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2715643Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2715680Z graph_break [] 2025-12-04T09:49:26.2715752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2715794Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2715848Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2715921Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2716149Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2716185Z graph_break [] 2025-12-04T09:49:26.2716257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2716297Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2716351Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2716423Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2716654Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2716690Z graph_break [] 2025-12-04T09:49:26.2716816Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2716876Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2716932Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2717002Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2717229Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2717264Z graph_break [] 2025-12-04T09:49:26.2717338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2717377Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2717432Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2717505Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2717732Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2717769Z graph_break [] 2025-12-04T09:49:26.2717842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2717882Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2717937Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2718008Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2718235Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2718272Z graph_break [] 2025-12-04T09:49:26.2718374Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2718414Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2718470Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2718543Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2718770Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2718806Z graph_break [] 2025-12-04T09:49:26.2718878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2718918Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2718973Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2719044Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2719273Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2719309Z graph_break [] 2025-12-04T09:49:26.2719384Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2719424Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2719478Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2719549Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2719776Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2719813Z graph_break [] 2025-12-04T09:49:26.2719885Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2719925Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2719981Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2720053Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2720295Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2720344Z graph_break [] 2025-12-04T09:49:26.2720417Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2720458Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2720512Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2720583Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2720811Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2720849Z graph_break [] 2025-12-04T09:49:26.2720921Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2720963Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2721018Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2721089Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2721316Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2721354Z graph_break [] 2025-12-04T09:49:26.2721426Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2721468Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2721522Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2721594Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2721842Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2721882Z graph_break [] 2025-12-04T09:49:26.2721954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2721996Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2722050Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2722122Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2722350Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2722388Z graph_break [] 2025-12-04T09:49:26.2722460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2722504Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2722560Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2722632Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2722860Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2722898Z graph_break [] 2025-12-04T09:49:26.2722970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2723012Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2723065Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2723137Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2723365Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2723411Z graph_break [] 2025-12-04T09:49:26.2723485Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2723535Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2723591Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2723662Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2723888Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2723924Z graph_break [] 2025-12-04T09:49:26.2723998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2724038Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2724094Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2724166Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2724397Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2724434Z graph_break [] 2025-12-04T09:49:26.2724507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2724548Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2724603Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2724674Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2724901Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2724936Z graph_break [] 2025-12-04T09:49:26.2725033Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2725075Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2725130Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2725201Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2725427Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2725463Z graph_break [] 2025-12-04T09:49:26.2725537Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2725577Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2725632Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2725702Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2725930Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2725968Z graph_break [] 2025-12-04T09:49:26.2726040Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2726080Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2726136Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2726206Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2726435Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2726471Z graph_break [] 2025-12-04T09:49:26.2726545Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2726586Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2726642Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2726726Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2727004Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2727041Z graph_break [] 2025-12-04T09:49:26.2727113Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2727154Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2727208Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2727280Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2727510Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2727547Z graph_break [] 2025-12-04T09:49:26.2727619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2727662Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2727716Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2727787Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2728014Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2728050Z graph_break [] 2025-12-04T09:49:26.2728134Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2728181Z Traceback (most recent call last): 2025-12-04T09:49:26.2728301Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2728373Z return value(self) 2025-12-04T09:49:26.2728521Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2728574Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2728711Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2728773Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2728931Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2729003Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2729056Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2729058Z 2025-12-04T09:49:26.2729110Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2729229Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2729329Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2729331Z 2025-12-04T09:49:26.2729405Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2729557Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2729559Z 2025-12-04T09:49:26.2729645Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2729720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2729761Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2729817Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2730048Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2730120Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2730172Z graph_break [] 2025-12-04T09:49:26.2730257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2730299Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2730353Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2730426Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2730655Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2730692Z graph_break [] 2025-12-04T09:49:26.2730764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2730805Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2730861Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2730934Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2731161Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2731199Z graph_break [] 2025-12-04T09:49:26.2731270Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2731311Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2731365Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2731437Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2731663Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2731724Z graph_break [] 2025-12-04T09:49:26.2731796Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2731841Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2731896Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2731968Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2732194Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2732231Z graph_break [] 2025-12-04T09:49:26.2732303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2732345Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2732398Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2732472Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2732702Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2732741Z graph_break [] 2025-12-04T09:49:26.2732813Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2732854Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2732908Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2732980Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2733208Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2733245Z graph_break [] 2025-12-04T09:49:26.2733317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2733360Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2733426Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2733498Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2733736Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2733771Z graph_break [] 2025-12-04T09:49:26.2733845Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2733885Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2733939Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2734010Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2734240Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2734276Z graph_break [] 2025-12-04T09:49:26.2734350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2734391Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2734445Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2734516Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2734744Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2734780Z graph_break [] 2025-12-04T09:49:26.2734853Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2734894Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2734966Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2735038Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2735266Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2735303Z graph_break [] 2025-12-04T09:49:26.2735376Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2735415Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2735471Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2735542Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2735768Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2735803Z graph_break [] 2025-12-04T09:49:26.2735880Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2735922Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2735980Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2736051Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2736280Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2736317Z graph_break [] 2025-12-04T09:49:26.2736393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2736433Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2736489Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2736561Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2736950Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2737018Z graph_break [] 2025-12-04T09:49:26.2737093Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2737134Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2737191Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2737263Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2737495Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2737532Z graph_break [] 2025-12-04T09:49:26.2737606Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2737649Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2737704Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2737780Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2738011Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2738049Z graph_break [] 2025-12-04T09:49:26.2738122Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2738165Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2738219Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2738292Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2738559Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2738600Z graph_break [] 2025-12-04T09:49:26.2738674Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2738719Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2738774Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2738848Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2739076Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2739115Z graph_break [] 2025-12-04T09:49:26.2739188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2739230Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2739284Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2739357Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2739586Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2739628Z graph_break [] 2025-12-04T09:49:26.2739700Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2739742Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2739796Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2739868Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2740094Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2740132Z graph_break [] 2025-12-04T09:49:26.2740208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2740263Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2740317Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2740402Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2740632Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2740670Z graph_break [] 2025-12-04T09:49:26.2740743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2740786Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2740842Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2740915Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2741145Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2741184Z graph_break [] 2025-12-04T09:49:26.2741258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2741300Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2741353Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2741426Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2741655Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2741692Z graph_break [] 2025-12-04T09:49:26.2741767Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2741807Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2741881Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2741955Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2742185Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2742220Z graph_break [] 2025-12-04T09:49:26.2742293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2742333Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2742389Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2742460Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2742693Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2742730Z graph_break [] 2025-12-04T09:49:26.2742806Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2742849Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2742906Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2742977Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2743206Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2743244Z graph_break [] 2025-12-04T09:49:26.2743318Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2743359Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2743414Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2743489Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2743727Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2743779Z graph_break [] 2025-12-04T09:49:26.2743853Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2743894Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2743949Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2744022Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2744249Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2744285Z graph_break [] 2025-12-04T09:49:26.2744359Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2744401Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2744458Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2744532Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2744765Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2744801Z graph_break [] 2025-12-04T09:49:26.2744875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2744917Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2744972Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2745044Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2745294Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2745335Z graph_break [] 2025-12-04T09:49:26.2745411Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2745456Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2745509Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2745581Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2745809Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2745849Z graph_break [] 2025-12-04T09:49:26.2745922Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2745964Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2746021Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2746096Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2746324Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2746364Z graph_break [] 2025-12-04T09:49:26.2746438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2746480Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2746534Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2746610Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2746883Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2746922Z graph_break [] 2025-12-04T09:49:26.2747011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2747070Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2747125Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2747199Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2747426Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2747463Z graph_break [] 2025-12-04T09:49:26.2747536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2747579Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2747637Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2747713Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2747939Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2747978Z graph_break [] 2025-12-04T09:49:26.2748051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2748093Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2748147Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2748219Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2748445Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2748481Z graph_break [] 2025-12-04T09:49:26.2748579Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2748623Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2748679Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2748753Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2748981Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2749019Z graph_break [] 2025-12-04T09:49:26.2749105Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2749152Z Traceback (most recent call last): 2025-12-04T09:49:26.2749275Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2749315Z return value(self) 2025-12-04T09:49:26.2749463Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2749515Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2749657Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2749719Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2749881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2749954Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2750010Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2750012Z 2025-12-04T09:49:26.2750063Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2750182Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2750281Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2750283Z 2025-12-04T09:49:26.2750359Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2750520Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2750532Z 2025-12-04T09:49:26.2750620Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2750695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2750740Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2750798Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2751030Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2751104Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2751140Z graph_break [] 2025-12-04T09:49:26.2751215Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2751260Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2751318Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2751390Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2751618Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2751658Z graph_break [] 2025-12-04T09:49:26.2751734Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2751777Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2751833Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2751907Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2752157Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2752197Z graph_break [] 2025-12-04T09:49:26.2752273Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2752316Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2752372Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2752444Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2752673Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2752709Z graph_break [] 2025-12-04T09:49:26.2752785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2752828Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2752885Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2752957Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2753187Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2753224Z graph_break [] 2025-12-04T09:49:26.2753300Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2753341Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2753397Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2753469Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2753700Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2753747Z graph_break [] 2025-12-04T09:49:26.2753823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2753873Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2753929Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2753999Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2754227Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2754264Z graph_break [] 2025-12-04T09:49:26.2754339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2754380Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2754436Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2754509Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2754741Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2754779Z graph_break [] 2025-12-04T09:49:26.2754854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2754895Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2754953Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2755026Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2755254Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2755291Z graph_break [] 2025-12-04T09:49:26.2755389Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2755432Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2755489Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2755562Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2755789Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2755829Z graph_break [] 2025-12-04T09:49:26.2755902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2755945Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2756000Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2756073Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2756302Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2756344Z graph_break [] 2025-12-04T09:49:26.2756417Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2756458Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2756512Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2756586Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2756858Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2756897Z graph_break [] 2025-12-04T09:49:26.2756970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2757014Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2757069Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2757172Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2757411Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2757448Z graph_break [] 2025-12-04T09:49:26.2757521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2757562Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2757616Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2757688Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2757917Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2757955Z graph_break [] 2025-12-04T09:49:26.2758028Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2758072Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2758127Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2758200Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2758426Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2758466Z graph_break [] 2025-12-04T09:49:26.2758540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2758583Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2758637Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2758734Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2758964Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2759004Z graph_break [] 2025-12-04T09:49:26.2759078Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2759119Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2759172Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2759244Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2759472Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2759508Z graph_break [] 2025-12-04T09:49:26.2759583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2759625Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2759682Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2759756Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2759983Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2760019Z graph_break [] 2025-12-04T09:49:26.2760093Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2760134Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2760188Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2760259Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2760490Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2760535Z graph_break [] 2025-12-04T09:49:26.2760619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2760660Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2760716Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2760786Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2761016Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2761052Z graph_break [] 2025-12-04T09:49:26.2761125Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2761165Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2761222Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2761293Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2761522Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2761559Z graph_break [] 2025-12-04T09:49:26.2761632Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2761673Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2761728Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2761800Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2762029Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2762084Z graph_break [] 2025-12-04T09:49:26.2762159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2762201Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2762256Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2762327Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2762555Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2762590Z graph_break [] 2025-12-04T09:49:26.2762664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2762705Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2762761Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2762831Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2763062Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2763100Z graph_break [] 2025-12-04T09:49:26.2763173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2763214Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2763270Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2763342Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2763570Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2763607Z graph_break [] 2025-12-04T09:49:26.2763680Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2763723Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2763790Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2763863Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2764101Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2764138Z graph_break [] 2025-12-04T09:49:26.2764211Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2764253Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2764307Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2764378Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2764606Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2764643Z graph_break [] 2025-12-04T09:49:26.2764716Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2764760Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2764814Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2764887Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2765114Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2765151Z graph_break [] 2025-12-04T09:49:26.2765223Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2765264Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2765318Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2765416Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2765644Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2765681Z graph_break [] 2025-12-04T09:49:26.2765753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2765795Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2765849Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2765921Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2766147Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2766184Z graph_break [] 2025-12-04T09:49:26.2766257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2766300Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2766355Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2766428Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2766654Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2766691Z graph_break [] 2025-12-04T09:49:26.2766805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2766845Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2766901Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2766972Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2767203Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2767275Z graph_break [] 2025-12-04T09:49:26.2767348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2767388Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2767443Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2767514Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2767741Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2767777Z graph_break [] 2025-12-04T09:49:26.2767850Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2767893Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2767948Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2768023Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2768252Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2768288Z graph_break [] 2025-12-04T09:49:26.2768360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2768400Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2768454Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2768526Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2768784Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2768821Z graph_break [] 2025-12-04T09:49:26.2768894Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2768936Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2768990Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2769060Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2769290Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2769326Z graph_break [] 2025-12-04T09:49:26.2769399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2769439Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2769494Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2769567Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2769796Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2769833Z graph_break [] 2025-12-04T09:49:26.2769906Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2769946Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2770001Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2770071Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2770300Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2770336Z graph_break [] 2025-12-04T09:49:26.2770424Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2770483Z Traceback (most recent call last): 2025-12-04T09:49:26.2770604Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2770655Z return value(self) 2025-12-04T09:49:26.2770801Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2770852Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2770990Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2771050Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2771209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2771282Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2771335Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2771338Z 2025-12-04T09:49:26.2771390Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2771509Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2771607Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2771609Z 2025-12-04T09:49:26.2771682Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2771832Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2771834Z 2025-12-04T09:49:26.2771919Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2771993Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2772033Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2772109Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2772338Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2772413Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2772449Z graph_break [] 2025-12-04T09:49:26.2772523Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2772564Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2772619Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2772690Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2772923Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2772959Z graph_break [] 2025-12-04T09:49:26.2773034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2773075Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2773131Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2773201Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2773429Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2773467Z graph_break [] 2025-12-04T09:49:26.2773539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2773580Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2773634Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2773706Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2773935Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2773993Z graph_break [] 2025-12-04T09:49:26.2774065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2774105Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2774160Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2774231Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2774458Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2774496Z graph_break [] 2025-12-04T09:49:26.2774569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2774611Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2774665Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2774739Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2774972Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2775008Z graph_break [] 2025-12-04T09:49:26.2775080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2775121Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2775175Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2775247Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2775492Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2775530Z graph_break [] 2025-12-04T09:49:26.2775602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2775644Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2775697Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2775769Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2775995Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2776033Z graph_break [] 2025-12-04T09:49:26.2776105Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2776145Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2776200Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2776273Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2776502Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2776541Z graph_break [] 2025-12-04T09:49:26.2776613Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2776654Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2776708Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2776812Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2777043Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2777081Z graph_break [] 2025-12-04T09:49:26.2777154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2777210Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2777278Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2777349Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2777576Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2777611Z graph_break [] 2025-12-04T09:49:26.2777684Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2777724Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2777779Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2777850Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2778078Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2778115Z graph_break [] 2025-12-04T09:49:26.2778189Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2778228Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2778283Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2778354Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2778580Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2778616Z graph_break [] 2025-12-04T09:49:26.2778688Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2778756Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2778812Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2778884Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2779117Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2779152Z graph_break [] 2025-12-04T09:49:26.2779225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2779265Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2779320Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2779390Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2779619Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2779656Z graph_break [] 2025-12-04T09:49:26.2779729Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2779770Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2779825Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2779896Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2780126Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2785037Z graph_break [] 2025-12-04T09:49:26.2785150Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2785197Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2785263Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2785345Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2785617Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2785671Z graph_break [] 2025-12-04T09:49:26.2785753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2785798Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2785857Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2785934Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2786167Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2786206Z graph_break [] 2025-12-04T09:49:26.2786284Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2786330Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2786386Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2786463Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2786691Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2786730Z graph_break [] 2025-12-04T09:49:26.2786882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2786926Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2786980Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2787054Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2787341Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2787386Z graph_break [] 2025-12-04T09:49:26.2787460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2787503Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2787559Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2787633Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2787864Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2787902Z graph_break [] 2025-12-04T09:49:26.2787977Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2788023Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2788079Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2788156Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2788387Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2788424Z graph_break [] 2025-12-04T09:49:26.2788498Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2788540Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2788595Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2788668Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2788898Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2788964Z graph_break [] 2025-12-04T09:49:26.2789038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2789095Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2789150Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2789224Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2789454Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2789492Z graph_break [] 2025-12-04T09:49:26.2789565Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2789608Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2789663Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2789738Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2789971Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2790011Z graph_break [] 2025-12-04T09:49:26.2790084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2790128Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2790182Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2790256Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2790487Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2790524Z graph_break [] 2025-12-04T09:49:26.2790628Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2790672Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2790729Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2790803Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2791033Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2791070Z graph_break [] 2025-12-04T09:49:26.2791145Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2791187Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2791243Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2791315Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2791547Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2791584Z graph_break [] 2025-12-04T09:49:26.2791660Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2791702Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2791759Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2791833Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2792064Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2792102Z graph_break [] 2025-12-04T09:49:26.2792177Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2792219Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2792276Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2792363Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2792605Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2792648Z graph_break [] 2025-12-04T09:49:26.2792731Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2792773Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2792829Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2792902Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2793135Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2793172Z graph_break [] 2025-12-04T09:49:26.2793253Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2813868Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2813938Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2814012Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2814252Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2814288Z graph_break [] 2025-12-04T09:49:26.2814363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2814405Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2814461Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2814577Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2814810Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2814850Z graph_break [] 2025-12-04T09:49:26.2814923Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2814965Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2815019Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2815091Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2815321Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2815357Z graph_break [] 2025-12-04T09:49:26.2815432Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2815474Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2815529Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2815602Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2815830Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2815866Z graph_break [] 2025-12-04T09:49:26.2815938Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2815980Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2816034Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2816105Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2816336Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2816389Z graph_break [] 2025-12-04T09:49:26.2816480Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2816521Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2816574Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2816646Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2816920Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2816957Z graph_break [] 2025-12-04T09:49:26.2817029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2817071Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2817128Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2817200Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2817429Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2817467Z graph_break [] 2025-12-04T09:49:26.2817539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2817580Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2817633Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2817706Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2817933Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2818007Z graph_break [] 2025-12-04T09:49:26.2818093Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2818142Z Traceback (most recent call last): 2025-12-04T09:49:26.2818267Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2818307Z return value(self) 2025-12-04T09:49:26.2818455Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2818508Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2818652Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2818714Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2818877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2818951Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2819010Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2819013Z 2025-12-04T09:49:26.2819065Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2819185Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2819284Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2819286Z 2025-12-04T09:49:26.2819360Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2819513Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2819515Z 2025-12-04T09:49:26.2819603Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2819676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2819721Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2819777Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2820024Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2820112Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2820149Z graph_break [] 2025-12-04T09:49:26.2820221Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2820264Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2820318Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2820390Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2820622Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2820660Z graph_break [] 2025-12-04T09:49:26.2820733Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2820777Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2820831Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2820903Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2821129Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2821166Z graph_break [] 2025-12-04T09:49:26.2821239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2821281Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2821335Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2821430Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2821659Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2821696Z graph_break [] 2025-12-04T09:49:26.2821769Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2821810Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2821864Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2821936Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2822162Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2822200Z graph_break [] 2025-12-04T09:49:26.2822274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2822317Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2822370Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2822443Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2822671Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2822708Z graph_break [] 2025-12-04T09:49:26.2822781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2822825Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2822880Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2822953Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2823186Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2823246Z graph_break [] 2025-12-04T09:49:26.2823318Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2823360Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2823414Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2823486Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2823714Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2823751Z graph_break [] 2025-12-04T09:49:26.2823823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2823865Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2823920Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2823994Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2824223Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2824260Z graph_break [] 2025-12-04T09:49:26.2824332Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2824373Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2824427Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2824500Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2824750Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2824788Z graph_break [] 2025-12-04T09:49:26.2824864Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2824906Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2824962Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2825034Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2825265Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2825302Z graph_break [] 2025-12-04T09:49:26.2825378Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2825418Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2825473Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2825546Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2825776Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2825814Z graph_break [] 2025-12-04T09:49:26.2825887Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2825928Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2825982Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2826053Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2826281Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2826317Z graph_break [] 2025-12-04T09:49:26.2826392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2826444Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2826501Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2826584Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2826861Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2826897Z graph_break [] 2025-12-04T09:49:26.2826971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2827011Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2827066Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2827136Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2827367Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2827404Z graph_break [] 2025-12-04T09:49:26.2827480Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2827520Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2827576Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2827648Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2827875Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2827911Z graph_break [] 2025-12-04T09:49:26.2827984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2828024Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2828109Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2828182Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2828410Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2828448Z graph_break [] 2025-12-04T09:49:26.2828521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2828562Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2828617Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2828689Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2828922Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2828961Z graph_break [] 2025-12-04T09:49:26.2829034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2829079Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2829134Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2829206Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2829435Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2829473Z graph_break [] 2025-12-04T09:49:26.2829545Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2829589Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2829643Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2829716Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2829947Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2830024Z graph_break [] 2025-12-04T09:49:26.2830097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2830140Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2830196Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2830269Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2830496Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2830534Z graph_break [] 2025-12-04T09:49:26.2830607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2830651Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2830706Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2830781Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2831012Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2831050Z graph_break [] 2025-12-04T09:49:26.2831124Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2831168Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2831223Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2831296Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2831548Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2831587Z graph_break [] 2025-12-04T09:49:26.2831660Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2831705Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2831760Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2831834Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2832061Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2832098Z graph_break [] 2025-12-04T09:49:26.2832171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2832212Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2832271Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2832342Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2832575Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2832612Z graph_break [] 2025-12-04T09:49:26.2832685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2832727Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2832782Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2832854Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2833085Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2833122Z graph_break [] 2025-12-04T09:49:26.2833199Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2833252Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2833319Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2833390Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2833622Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2833657Z graph_break [] 2025-12-04T09:49:26.2833732Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2833773Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2833828Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2833899Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2834131Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2834169Z graph_break [] 2025-12-04T09:49:26.2834243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2834284Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2834339Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2834409Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2834641Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2834677Z graph_break [] 2025-12-04T09:49:26.2834752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2834813Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2834870Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2834944Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2835178Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2835214Z graph_break [] 2025-12-04T09:49:26.2835288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2835329Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2835386Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2835459Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2835690Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2835728Z graph_break [] 2025-12-04T09:49:26.2835804Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2835847Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2835905Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2835977Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2836208Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2836246Z graph_break [] 2025-12-04T09:49:26.2836319Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2836363Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2836416Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2836495Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2836777Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2836842Z graph_break [] 2025-12-04T09:49:26.2836917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2836966Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2837019Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2837094Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2837324Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2837362Z graph_break [] 2025-12-04T09:49:26.2837438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2837482Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2837537Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2837612Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2837842Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2837879Z graph_break [] 2025-12-04T09:49:26.2837953Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2837996Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2838052Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2838127Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2838388Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2838435Z graph_break [] 2025-12-04T09:49:26.2838508Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2838553Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2838610Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2838685Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2838916Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2838956Z graph_break [] 2025-12-04T09:49:26.2839029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2839074Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2839132Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2839206Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2839436Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2839475Z graph_break [] 2025-12-04T09:49:26.2839549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2839592Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2839648Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2839723Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2839955Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2839993Z graph_break [] 2025-12-04T09:49:26.2840082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2840136Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2840190Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2840265Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2840494Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2840535Z graph_break [] 2025-12-04T09:49:26.2840622Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2840669Z Traceback (most recent call last): 2025-12-04T09:49:26.2840791Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2840836Z return value(self) 2025-12-04T09:49:26.2840985Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2841041Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2841183Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2841245Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2841407Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2841480Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2841537Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2841539Z 2025-12-04T09:49:26.2841590Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2841734Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2841833Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2841837Z 2025-12-04T09:49:26.2841915Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2842067Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2842069Z 2025-12-04T09:49:26.2842160Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2842235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2842278Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2842336Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2842571Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2842645Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2842686Z graph_break [] 2025-12-04T09:49:26.2842761Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2842805Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2842861Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2842934Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2843165Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2843205Z graph_break [] 2025-12-04T09:49:26.2843278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2843322Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2843378Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2843460Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2843710Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2843760Z graph_break [] 2025-12-04T09:49:26.2843832Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2843874Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2843928Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2844001Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2844231Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2844271Z graph_break [] 2025-12-04T09:49:26.2844346Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2844392Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2844449Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2844522Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2844750Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2844790Z graph_break [] 2025-12-04T09:49:26.2844865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2844908Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2844964Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2845037Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2845291Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2845331Z graph_break [] 2025-12-04T09:49:26.2845403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2845444Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2845500Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2845574Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2845803Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2845841Z graph_break [] 2025-12-04T09:49:26.2845914Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2845961Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2846017Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2846091Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2846321Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2846361Z graph_break [] 2025-12-04T09:49:26.2846434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2846477Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2846531Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2846604Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2846874Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2846935Z graph_break [] 2025-12-04T09:49:26.2847013Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2847088Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2847145Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2847217Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2847447Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2847486Z graph_break [] 2025-12-04T09:49:26.2847562Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2847604Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2847661Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2847735Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2847969Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2848007Z graph_break [] 2025-12-04T09:49:26.2848082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2848124Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2848181Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2848253Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2848482Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2848519Z graph_break [] 2025-12-04T09:49:26.2848641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2848685Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2848746Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2848818Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2849049Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2849087Z graph_break [] 2025-12-04T09:49:26.2849164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2849205Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2849262Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2849334Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2849567Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2849605Z graph_break [] 2025-12-04T09:49:26.2849678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2849719Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2849776Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2849848Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2850078Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2850113Z graph_break [] 2025-12-04T09:49:26.2850187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2850230Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2850285Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2850373Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2850615Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2850652Z graph_break [] 2025-12-04T09:49:26.2850727Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2850770Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2850825Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2850898Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2851127Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2851167Z graph_break [] 2025-12-04T09:49:26.2851241Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2851287Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2851342Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2851415Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2851643Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2851682Z graph_break [] 2025-12-04T09:49:26.2851756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2851798Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2851853Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2851946Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2852177Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2852215Z graph_break [] 2025-12-04T09:49:26.2852287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2852330Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2852386Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2852459Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2852688Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2852725Z graph_break [] 2025-12-04T09:49:26.2852799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2852843Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2852897Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2852971Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2853199Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2853238Z graph_break [] 2025-12-04T09:49:26.2853312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2853354Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2853408Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2853480Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2853709Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2853768Z graph_break [] 2025-12-04T09:49:26.2853841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2853884Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2853938Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2854010Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2854238Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2854275Z graph_break [] 2025-12-04T09:49:26.2854348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2854389Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2854445Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2854517Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2854748Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2854785Z graph_break [] 2025-12-04T09:49:26.2854860Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2854901Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2854959Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2855031Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2855285Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2855322Z graph_break [] 2025-12-04T09:49:26.2855398Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2855440Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2855496Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2855567Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2855799Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2855836Z graph_break [] 2025-12-04T09:49:26.2855909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2855950Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2856005Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2856078Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2856309Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2856347Z graph_break [] 2025-12-04T09:49:26.2856421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2856462Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2856518Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2856589Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2856853Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2856889Z graph_break [] 2025-12-04T09:49:26.2856966Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2857007Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2857089Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2857175Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2857405Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2857442Z graph_break [] 2025-12-04T09:49:26.2857516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2857556Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2857611Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2857684Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2857914Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2857952Z graph_break [] 2025-12-04T09:49:26.2858026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2858068Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2858123Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2858195Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2858422Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2858459Z graph_break [] 2025-12-04T09:49:26.2858531Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2858572Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2858655Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2858728Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2858957Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2858995Z graph_break [] 2025-12-04T09:49:26.2859067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2859108Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2859162Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2859235Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2859465Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2859503Z graph_break [] 2025-12-04T09:49:26.2859576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2859619Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2859674Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2859746Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2859974Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2860011Z graph_break [] 2025-12-04T09:49:26.2860083Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2860125Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2860179Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2860250Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2860481Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2860540Z graph_break [] 2025-12-04T09:49:26.2860612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2860654Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2860708Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2860779Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2861007Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2861044Z graph_break [] 2025-12-04T09:49:26.2861116Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2861159Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2861214Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2861286Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2861518Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2861555Z graph_break [] 2025-12-04T09:49:26.2861628Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2861670Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2861723Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2861795Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2862041Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2862079Z graph_break [] 2025-12-04T09:49:26.2862153Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2862195Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2862250Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2862321Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2862551Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2862586Z graph_break [] 2025-12-04T09:49:26.2862660Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2862700Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2862754Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2862827Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2863058Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2863095Z graph_break [] 2025-12-04T09:49:26.2863169Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2863210Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2863264Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2863336Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2863568Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2863604Z graph_break [] 2025-12-04T09:49:26.2863692Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2863748Z Traceback (most recent call last): 2025-12-04T09:49:26.2863879Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2863919Z return value(self) 2025-12-04T09:49:26.2864064Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2864115Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2864254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2864314Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2864475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2864548Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2864603Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2864606Z 2025-12-04T09:49:26.2864658Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2864775Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2864873Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2864877Z 2025-12-04T09:49:26.2864950Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2865103Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2865105Z 2025-12-04T09:49:26.2865191Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2865265Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2865307Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2865382Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2865614Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2865689Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2865725Z graph_break [] 2025-12-04T09:49:26.2865798Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2865839Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2865894Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2865964Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2866196Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2866233Z graph_break [] 2025-12-04T09:49:26.2866308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2866349Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2866405Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2866476Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2866705Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2866787Z graph_break [] 2025-12-04T09:49:26.2866861Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2866902Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2866957Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2867029Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2867275Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2867332Z graph_break [] 2025-12-04T09:49:26.2867408Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2867448Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2867503Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2867575Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2867806Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2867842Z graph_break [] 2025-12-04T09:49:26.2867917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2867958Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2868012Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2868083Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2868314Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2868350Z graph_break [] 2025-12-04T09:49:26.2868422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2868464Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2868517Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2868588Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2868858Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2868897Z graph_break [] 2025-12-04T09:49:26.2868970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2869012Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2869065Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2869137Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2869364Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2869401Z graph_break [] 2025-12-04T09:49:26.2869474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2869515Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2869571Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2869644Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2869871Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2869909Z graph_break [] 2025-12-04T09:49:26.2869981Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2870023Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2870076Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2870148Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2870374Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2870412Z graph_break [] 2025-12-04T09:49:26.2870495Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2870536Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2870602Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2870673Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2870900Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2870937Z graph_break [] 2025-12-04T09:49:26.2871009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2871051Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2871106Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2871181Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2871409Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2871448Z graph_break [] 2025-12-04T09:49:26.2871520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2871561Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2871615Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2871687Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2871914Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2871951Z graph_break [] 2025-12-04T09:49:26.2872042Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2872084Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2872139Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2872212Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2872439Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2872474Z graph_break [] 2025-12-04T09:49:26.2872548Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2872588Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2872643Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2872713Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2872944Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2872982Z graph_break [] 2025-12-04T09:49:26.2873056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2873097Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2873153Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2873224Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2873453Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2873489Z graph_break [] 2025-12-04T09:49:26.2873563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2873603Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2873659Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2873730Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2873969Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2874015Z graph_break [] 2025-12-04T09:49:26.2874089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2874130Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2874185Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2874256Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2874483Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2874521Z graph_break [] 2025-12-04T09:49:26.2874595Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2874637Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2874694Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2874765Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2874993Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2875029Z graph_break [] 2025-12-04T09:49:26.2875103Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2875143Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2875199Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2875270Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2875519Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2875558Z graph_break [] 2025-12-04T09:49:26.2875632Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2875673Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2875728Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2875799Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2876026Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2876063Z graph_break [] 2025-12-04T09:49:26.2876135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2876178Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2876234Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2876306Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2876533Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2876569Z graph_break [] 2025-12-04T09:49:26.2876641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2876683Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2876736Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2877079Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2877309Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2877366Z graph_break [] 2025-12-04T09:49:26.2877438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2877495Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2877549Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2877622Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2877850Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2877886Z graph_break [] 2025-12-04T09:49:26.2877958Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2877999Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2878052Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2878126Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2878357Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2878395Z graph_break [] 2025-12-04T09:49:26.2878467Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2878509Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2878562Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2878636Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2878865Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2878903Z graph_break [] 2025-12-04T09:49:26.2879001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2879046Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2879101Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2879176Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2879403Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2879441Z graph_break [] 2025-12-04T09:49:26.2879513Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2879555Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2879609Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2879681Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2879911Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2879948Z graph_break [] 2025-12-04T09:49:26.2880022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2880062Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2880117Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2880189Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2880417Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2880453Z graph_break [] 2025-12-04T09:49:26.2880526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2880568Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2880623Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2880706Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2880949Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2880985Z graph_break [] 2025-12-04T09:49:26.2881059Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2881100Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2881156Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2881227Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2881456Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2881492Z graph_break [] 2025-12-04T09:49:26.2881566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2881608Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2881663Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2881735Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2881964Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2882000Z graph_break [] 2025-12-04T09:49:26.2882074Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2882115Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2882171Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2882266Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2882498Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2882535Z graph_break [] 2025-12-04T09:49:26.2882609Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2882649Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2882705Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2882776Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2883006Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2883041Z graph_break [] 2025-12-04T09:49:26.2883116Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2883157Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2883213Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2883285Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2883514Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2883550Z graph_break [] 2025-12-04T09:49:26.2883624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2883667Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2883721Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2883793Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2884023Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2884071Z graph_break [] 2025-12-04T09:49:26.2884154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2884197Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2884252Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2884324Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2884553Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2884589Z graph_break [] 2025-12-04T09:49:26.2884661Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2884702Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2884757Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2884830Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2885059Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2885096Z graph_break [] 2025-12-04T09:49:26.2885168Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2885209Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2885262Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2885334Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2885582Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2885618Z graph_break [] 2025-12-04T09:49:26.2885692Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2885734Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2885788Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2885860Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2886087Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2886124Z graph_break [] 2025-12-04T09:49:26.2886196Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2886237Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2886291Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2886365Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2886593Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2886633Z graph_break [] 2025-12-04T09:49:26.2886704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2886775Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2886829Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2886901Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2887128Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2887164Z graph_break [] 2025-12-04T09:49:26.2887251Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2887298Z Traceback (most recent call last): 2025-12-04T09:49:26.2887442Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2887495Z return value(self) 2025-12-04T09:49:26.2887641Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2887692Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2887832Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2887893Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2888055Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2888127Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2888185Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2888187Z 2025-12-04T09:49:26.2888241Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2888358Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2888461Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2888463Z 2025-12-04T09:49:26.2888538Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2888688Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2888691Z 2025-12-04T09:49:26.2888780Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2888854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2888897Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2888954Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2889211Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2889287Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2889326Z graph_break [] 2025-12-04T09:49:26.2889399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2889444Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2889501Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2889575Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2889809Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2889849Z graph_break [] 2025-12-04T09:49:26.2889924Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2889968Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2890026Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2890099Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2890330Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2890369Z graph_break [] 2025-12-04T09:49:26.2890444Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2890486Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2890545Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2890617Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2890850Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2890914Z graph_break [] 2025-12-04T09:49:26.2890990Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2891032Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2891090Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2891162Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2891392Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2891432Z graph_break [] 2025-12-04T09:49:26.2891508Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2891551Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2891608Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2891681Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2891912Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2891949Z graph_break [] 2025-12-04T09:49:26.2892025Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2892066Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2892124Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2892194Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2892442Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2892481Z graph_break [] 2025-12-04T09:49:26.2892555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2892597Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2892655Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2892728Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2892958Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2892994Z graph_break [] 2025-12-04T09:49:26.2893070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2893112Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2893169Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2893242Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2893480Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2893518Z graph_break [] 2025-12-04T09:49:26.2893597Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2893637Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2893693Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2893766Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2894001Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2894043Z graph_break [] 2025-12-04T09:49:26.2894117Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2894175Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2894228Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2894311Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2894541Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2894581Z graph_break [] 2025-12-04T09:49:26.2894654Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2894699Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2894754Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2894828Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2895058Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2895101Z graph_break [] 2025-12-04T09:49:26.2895173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2895217Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2895271Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2895346Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2895575Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2895613Z graph_break [] 2025-12-04T09:49:26.2895686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2895728Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2895801Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2895876Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2896106Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2896145Z graph_break [] 2025-12-04T09:49:26.2896218Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2896260Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2896315Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2896387Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2896619Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2896656Z graph_break [] 2025-12-04T09:49:26.2896730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2896812Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2896866Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2896941Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2897170Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2897208Z graph_break [] 2025-12-04T09:49:26.2897280Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2897326Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2897382Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2897455Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2897683Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2897759Z graph_break [] 2025-12-04T09:49:26.2897834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2897877Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2897932Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2898006Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2898238Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2898275Z graph_break [] 2025-12-04T09:49:26.2898354Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2898398Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2898456Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2898529Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2898761Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2898797Z graph_break [] 2025-12-04T09:49:26.2898872Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2898915Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2898973Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2899047Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2899304Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2899343Z graph_break [] 2025-12-04T09:49:26.2899419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2899460Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2899518Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2899591Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2899821Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2899858Z graph_break [] 2025-12-04T09:49:26.2899933Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2899974Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2900033Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2900105Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2900335Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2900373Z graph_break [] 2025-12-04T09:49:26.2900447Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2900489Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2900547Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2900619Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2900851Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2900889Z graph_break [] 2025-12-04T09:49:26.2900974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2901016Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2901084Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2901157Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2901388Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2901425Z graph_break [] 2025-12-04T09:49:26.2901503Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2901547Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2901604Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2901678Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2901912Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2901952Z graph_break [] 2025-12-04T09:49:26.2902025Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2902069Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2902124Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2902199Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2902426Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2902465Z graph_break [] 2025-12-04T09:49:26.2902557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2902602Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2902657Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2902730Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2902959Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2902996Z graph_break [] 2025-12-04T09:49:26.2903069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2903111Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2903165Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2903238Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2903466Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2903504Z graph_break [] 2025-12-04T09:49:26.2903577Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2903623Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2903678Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2903752Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2903981Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2904020Z graph_break [] 2025-12-04T09:49:26.2904092Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2904134Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2904191Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2904265Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2904504Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2904552Z graph_break [] 2025-12-04T09:49:26.2904625Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2904667Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2904722Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2904796Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2905024Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2905062Z graph_break [] 2025-12-04T09:49:26.2905136Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2905178Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2905234Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2905306Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2905534Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2905571Z graph_break [] 2025-12-04T09:49:26.2905648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2905689Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2905744Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2905815Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2906071Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2906110Z graph_break [] 2025-12-04T09:49:26.2906184Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2906225Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2906281Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2906353Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2906581Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2906617Z graph_break [] 2025-12-04T09:49:26.2906690Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2906732Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2906824Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2906895Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2907126Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2907162Z graph_break [] 2025-12-04T09:49:26.2907235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2907275Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2907331Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2907402Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2907631Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2907689Z graph_break [] 2025-12-04T09:49:26.2907766Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2907821Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2907878Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2907949Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2908179Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2908216Z graph_break [] 2025-12-04T09:49:26.2908292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2908332Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2908390Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2908464Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2908696Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2908734Z graph_break [] 2025-12-04T09:49:26.2908807Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2908847Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2908905Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2908976Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2909204Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2909241Z graph_break [] 2025-12-04T09:49:26.2909337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2909381Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2909436Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2909509Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2909737Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2909775Z graph_break [] 2025-12-04T09:49:26.2909849Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2909891Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2909946Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2910018Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2910247Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2910286Z graph_break [] 2025-12-04T09:49:26.2910359Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2910401Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2910456Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2910529Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2910758Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2910795Z graph_break [] 2025-12-04T09:49:26.2910868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2910910Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2910966Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2911049Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2911287Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2911325Z graph_break [] 2025-12-04T09:49:26.2911411Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2911459Z Traceback (most recent call last): 2025-12-04T09:49:26.2911577Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2911618Z return value(self) 2025-12-04T09:49:26.2911763Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2911816Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2911954Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2912017Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2912176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2912249Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2912301Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2912304Z 2025-12-04T09:49:26.2912356Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2912471Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2912568Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2912570Z 2025-12-04T09:49:26.2912665Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2912815Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2912819Z 2025-12-04T09:49:26.2912905Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2912979Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2913023Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2913079Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2913310Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2913383Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2913420Z graph_break [] 2025-12-04T09:49:26.2913494Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2913537Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2913594Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2913668Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2913898Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2913935Z graph_break [] 2025-12-04T09:49:26.2914009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2914050Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2914105Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2914177Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2914406Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2914460Z graph_break [] 2025-12-04T09:49:26.2914545Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2914586Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2914640Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2914712Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2914939Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2914976Z graph_break [] 2025-12-04T09:49:26.2915048Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2915089Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2915145Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2915217Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2915445Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2915483Z graph_break [] 2025-12-04T09:49:26.2915555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2915596Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2915649Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2915721Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2915951Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2916005Z graph_break [] 2025-12-04T09:49:26.2916080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2916121Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2916178Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2916249Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2916479Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2916515Z graph_break [] 2025-12-04T09:49:26.2916588Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2916628Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2916682Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2916780Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2917011Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2917048Z graph_break [] 2025-12-04T09:49:26.2917122Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2917162Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2917216Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2917287Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2917517Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2917553Z graph_break [] 2025-12-04T09:49:26.2917627Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2917668Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2917741Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2917812Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2918056Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2918092Z graph_break [] 2025-12-04T09:49:26.2918166Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2918206Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2918263Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2918334Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2918566Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2918603Z graph_break [] 2025-12-04T09:49:26.2918676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2918718Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2918775Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2918846Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2919075Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2919111Z graph_break [] 2025-12-04T09:49:26.2919186Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2919227Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2919283Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2919395Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2919626Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2919664Z graph_break [] 2025-12-04T09:49:26.2919737Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2919778Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2919832Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2919904Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2920133Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2920172Z graph_break [] 2025-12-04T09:49:26.2920245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2920289Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2920345Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2920416Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2920642Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2920679Z graph_break [] 2025-12-04T09:49:26.2920750Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2920793Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2920847Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2920920Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2921150Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2921207Z graph_break [] 2025-12-04T09:49:26.2921280Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2921321Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2921376Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2921448Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2921675Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2921712Z graph_break [] 2025-12-04T09:49:26.2921785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2921828Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2921883Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2921958Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2922189Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2922227Z graph_break [] 2025-12-04T09:49:26.2922299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2922341Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2922395Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2922469Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2922716Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2922755Z graph_break [] 2025-12-04T09:49:26.2922828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2922871Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2922927Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2922999Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2923227Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2923264Z graph_break [] 2025-12-04T09:49:26.2923338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2923378Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2923434Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2923507Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2923736Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2923773Z graph_break [] 2025-12-04T09:49:26.2923847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2923888Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2923943Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2926030Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2926260Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2926297Z graph_break [] 2025-12-04T09:49:26.2926376Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2926437Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2926493Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2926576Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2926839Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2926875Z graph_break [] 2025-12-04T09:49:26.2926949Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2926989Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2927044Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2927116Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2927349Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2927389Z graph_break [] 2025-12-04T09:49:26.2927462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2927502Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2927556Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2927627Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2927856Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2927892Z graph_break [] 2025-12-04T09:49:26.2927967Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2928007Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2928096Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2928171Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2928401Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2928436Z graph_break [] 2025-12-04T09:49:26.2928510Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2928550Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2928605Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2928676Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2928907Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2928942Z graph_break [] 2025-12-04T09:49:26.2929019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2929060Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2929115Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2929187Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2929416Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2929452Z graph_break [] 2025-12-04T09:49:26.2929525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2929565Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2929620Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2929693Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2929921Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2929987Z graph_break [] 2025-12-04T09:49:26.2930060Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2930101Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2930155Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2930227Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2930453Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2930490Z graph_break [] 2025-12-04T09:49:26.2930564Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2930605Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2930660Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2930733Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2930960Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2930996Z graph_break [] 2025-12-04T09:49:26.2931069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2931110Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2931165Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2931237Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2931484Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2931523Z graph_break [] 2025-12-04T09:49:26.2931596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2931637Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2931691Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2931762Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2931990Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2932027Z graph_break [] 2025-12-04T09:49:26.2932099Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2932141Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2932196Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2932270Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2932497Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2932535Z graph_break [] 2025-12-04T09:49:26.2932607Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2932649Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2932703Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2932775Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2933001Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2933039Z graph_break [] 2025-12-04T09:49:26.2933111Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2933168Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2933239Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2933311Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2933540Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2933576Z graph_break [] 2025-12-04T09:49:26.2933649Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2933690Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2933745Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2933816Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2934049Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2934087Z graph_break [] 2025-12-04T09:49:26.2934161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2934201Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2934256Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2934328Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2934556Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2934592Z graph_break [] 2025-12-04T09:49:26.2934666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2934726Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2934784Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2934856Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2935089Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2935124Z graph_break [] 2025-12-04T09:49:26.2935197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2935237Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2935291Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2935362Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2935591Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2935628Z graph_break [] 2025-12-04T09:49:26.2935701Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2935742Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2935797Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2935869Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2936100Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2936135Z graph_break [] 2025-12-04T09:49:26.2936209Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2936249Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2936304Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2936377Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2936617Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2936663Z graph_break [] 2025-12-04T09:49:26.2936736Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2936808Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2936864Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2936936Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2937163Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2937200Z graph_break [] 2025-12-04T09:49:26.2937274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2937317Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2937373Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2937445Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2937672Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2937709Z graph_break [] 2025-12-04T09:49:26.2937794Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2937841Z Traceback (most recent call last): 2025-12-04T09:49:26.2937964Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2938004Z return value(self) 2025-12-04T09:49:26.2938185Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2938238Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2938380Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2938441Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2938602Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2938676Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2938729Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2938732Z 2025-12-04T09:49:26.2938785Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2938902Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2939004Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2939007Z 2025-12-04T09:49:26.2939080Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2939233Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2939236Z 2025-12-04T09:49:26.2939322Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2939397Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2939439Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2939496Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2939724Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2939798Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2939834Z graph_break [] 2025-12-04T09:49:26.2939924Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2939980Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2940036Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2940107Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2940336Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2940372Z graph_break [] 2025-12-04T09:49:26.2940445Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2940486Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2940540Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2940613Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2940843Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2940881Z graph_break [] 2025-12-04T09:49:26.2940955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2940997Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2941051Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2941123Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2941351Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2941387Z graph_break [] 2025-12-04T09:49:26.2941480Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2941524Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2941578Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2941653Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2941881Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2941918Z graph_break [] 2025-12-04T09:49:26.2941991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2942031Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2942084Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2942156Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2942384Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2942424Z graph_break [] 2025-12-04T09:49:26.2942498Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2942540Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2942593Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2942665Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2942892Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2942929Z graph_break [] 2025-12-04T09:49:26.2943001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2943042Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2943098Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2943183Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2943410Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2943458Z graph_break [] 2025-12-04T09:49:26.2943532Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2943574Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2943628Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2943700Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2943928Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2943966Z graph_break [] 2025-12-04T09:49:26.2944041Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2944082Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2944137Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2944208Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2944437Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2944473Z graph_break [] 2025-12-04T09:49:26.2944547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2944587Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2944642Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2944734Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2944963Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2945000Z graph_break [] 2025-12-04T09:49:26.2945075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2945116Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2945172Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2945243Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2945472Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2945508Z graph_break [] 2025-12-04T09:49:26.2945584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2945624Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2945680Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2945752Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2945983Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2946019Z graph_break [] 2025-12-04T09:49:26.2946094Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2946134Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2946189Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2946261Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2946492Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2946545Z graph_break [] 2025-12-04T09:49:26.2946619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2946671Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2946726Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2946859Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2947087Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2947123Z graph_break [] 2025-12-04T09:49:26.2947196Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2947236Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2947294Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2947366Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2947596Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2947636Z graph_break [] 2025-12-04T09:49:26.2947708Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2947752Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2947806Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2947881Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2948106Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2948169Z graph_break [] 2025-12-04T09:49:26.2948243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2948286Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2948343Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2948415Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2948644Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2948680Z graph_break [] 2025-12-04T09:49:26.2948754Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2948796Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2948852Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2948926Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2949156Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2949195Z graph_break [] 2025-12-04T09:49:26.2949267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2949308Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2949361Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2949433Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2949660Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2949696Z graph_break [] 2025-12-04T09:49:26.2949768Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2949811Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2949881Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2949954Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2950195Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2950236Z graph_break [] 2025-12-04T09:49:26.2950309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2950352Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2950406Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2950478Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2950713Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2950753Z graph_break [] 2025-12-04T09:49:26.2950825Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2950868Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2950922Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2950995Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2951231Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2951267Z graph_break [] 2025-12-04T09:49:26.2951340Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2951380Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2951437Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2951528Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2951760Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2951798Z graph_break [] 2025-12-04T09:49:26.2951875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2951915Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2951972Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2952042Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2952269Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2952306Z graph_break [] 2025-12-04T09:49:26.2952382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2952425Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2952480Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2952553Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2952784Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2952822Z graph_break [] 2025-12-04T09:49:26.2952894Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2952937Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2952991Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2953064Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2953293Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2953356Z graph_break [] 2025-12-04T09:49:26.2953428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2953470Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2953523Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2953596Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2953826Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2953862Z graph_break [] 2025-12-04T09:49:26.2953935Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2953982Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2954036Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2954110Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2954340Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2954380Z graph_break [] 2025-12-04T09:49:26.2954453Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2954497Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2954552Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2954624Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2954873Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2954914Z graph_break [] 2025-12-04T09:49:26.2954989Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2955034Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2955090Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2955163Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2955393Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2955430Z graph_break [] 2025-12-04T09:49:26.2955504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2955544Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2955598Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2955672Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2955901Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2955939Z graph_break [] 2025-12-04T09:49:26.2956011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2956053Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2956107Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2956179Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2956405Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2956442Z graph_break [] 2025-12-04T09:49:26.2956516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2956571Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2956625Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2956707Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2956991Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2957028Z graph_break [] 2025-12-04T09:49:26.2957100Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2957141Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2957196Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2957267Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2957495Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2957532Z graph_break [] 2025-12-04T09:49:26.2957605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2957647Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2957700Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2957772Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2957997Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2958035Z graph_break [] 2025-12-04T09:49:26.2958107Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2958148Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2958236Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2958310Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2958538Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2958575Z graph_break [] 2025-12-04T09:49:26.2958649Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2958689Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2958744Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2958815Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2959045Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2959080Z graph_break [] 2025-12-04T09:49:26.2959155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2959196Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2959250Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2959321Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2959550Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2959585Z graph_break [] 2025-12-04T09:49:26.2959658Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2959698Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2959753Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2959826Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2960054Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2960118Z graph_break [] 2025-12-04T09:49:26.2960192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2960232Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2960286Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2960357Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2960587Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2960622Z graph_break [] 2025-12-04T09:49:26.2960697Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2960737Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2960794Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2960867Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2961095Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2961131Z graph_break [] 2025-12-04T09:49:26.2961205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2961245Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2961300Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2961371Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2961618Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2961655Z graph_break [] 2025-12-04T09:49:26.2961730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2961770Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2961825Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2961896Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2962123Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2962160Z graph_break [] 2025-12-04T09:49:26.2962232Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2962272Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2962329Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2962401Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2962630Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2962668Z graph_break [] 2025-12-04T09:49:26.2962753Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2962799Z Traceback (most recent call last): 2025-12-04T09:49:26.2962919Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2962958Z return value(self) 2025-12-04T09:49:26.2963104Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2963155Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2963295Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2963373Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2963544Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2963616Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2963669Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2963671Z 2025-12-04T09:49:26.2963723Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2963839Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2963938Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2963940Z 2025-12-04T09:49:26.2964013Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2964166Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2964169Z 2025-12-04T09:49:26.2964256Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2964331Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2964373Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2964429Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2964658Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2964732Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2964767Z graph_break [] 2025-12-04T09:49:26.2964841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2964902Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2964958Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2965030Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2965259Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2965296Z graph_break [] 2025-12-04T09:49:26.2965368Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2965409Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2965463Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2965535Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2965765Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2965802Z graph_break [] 2025-12-04T09:49:26.2965875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2965917Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2965972Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2966043Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2966271Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2966307Z graph_break [] 2025-12-04T09:49:26.2966379Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2966420Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2966473Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2966547Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2966826Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2966887Z graph_break [] 2025-12-04T09:49:26.2966959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2967001Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2967055Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2967127Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2967353Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2967390Z graph_break [] 2025-12-04T09:49:26.2967463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2967505Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2967559Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2967632Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2967859Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2967895Z graph_break [] 2025-12-04T09:49:26.2967967Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2968008Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2968062Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2968134Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2968388Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2968427Z graph_break [] 2025-12-04T09:49:26.2968500Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2968541Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2968595Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2968668Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2968897Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2968934Z graph_break [] 2025-12-04T09:49:26.2969007Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2969047Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2969104Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2969176Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2969406Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2969442Z graph_break [] 2025-12-04T09:49:26.2969515Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2969555Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2969610Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2969681Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2969912Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2969948Z graph_break [] 2025-12-04T09:49:26.2970036Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2970087Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2970142Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2970213Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2970441Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2970477Z graph_break [] 2025-12-04T09:49:26.2970550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2970591Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2970645Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2970717Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2970946Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2970983Z graph_break [] 2025-12-04T09:49:26.2971056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2971096Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2971151Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2971222Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2971451Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2971487Z graph_break [] 2025-12-04T09:49:26.2971585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2971627Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2971682Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2971754Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2971984Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2972019Z graph_break [] 2025-12-04T09:49:26.2972092Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2972133Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2972188Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2972258Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2972489Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2972526Z graph_break [] 2025-12-04T09:49:26.2972601Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2972641Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2972697Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2972767Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2972997Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2973033Z graph_break [] 2025-12-04T09:49:26.2973106Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2973147Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2973202Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2973286Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2973513Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2973559Z graph_break [] 2025-12-04T09:49:26.2973632Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2973673Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2973727Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2973799Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2974028Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2974067Z graph_break [] 2025-12-04T09:49:26.2974141Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2974182Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2974237Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2974309Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2974537Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2974574Z graph_break [] 2025-12-04T09:49:26.2974647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2974688Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2974742Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2974814Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2975062Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2975101Z graph_break [] 2025-12-04T09:49:26.2975173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2975215Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2975268Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2975340Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2975567Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2975604Z graph_break [] 2025-12-04T09:49:26.2975678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2975720Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2975775Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2975847Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2976075Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2976111Z graph_break [] 2025-12-04T09:49:26.2976183Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2976224Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2976277Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2976349Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2976581Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2976630Z graph_break [] 2025-12-04T09:49:26.2976704Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2976814Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2976870Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2976941Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2977170Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2977206Z graph_break [] 2025-12-04T09:49:26.2977280Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2977320Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2977377Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2977448Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2977678Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2977715Z graph_break [] 2025-12-04T09:49:26.2977789Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2977830Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2977885Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2977957Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2978185Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2978259Z graph_break [] 2025-12-04T09:49:26.2978334Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2978376Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2978433Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2978503Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2978734Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2978770Z graph_break [] 2025-12-04T09:49:26.2978844Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2978885Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2978942Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2979012Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2979244Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2979282Z graph_break [] 2025-12-04T09:49:26.2979355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2979395Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2979450Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2979522Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2979750Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2979785Z graph_break [] 2025-12-04T09:49:26.2979858Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2979900Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2979972Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2980044Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2980290Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2980327Z graph_break [] 2025-12-04T09:49:26.2980400Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2980442Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2980497Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2980569Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2980801Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2980839Z graph_break [] 2025-12-04T09:49:26.2980911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2980956Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2981010Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2981083Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2981310Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2981348Z graph_break [] 2025-12-04T09:49:26.2981420Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2981462Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2981516Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2981609Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2981838Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2981877Z graph_break [] 2025-12-04T09:49:26.2981949Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2981989Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2982043Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2982114Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2982341Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2982378Z graph_break [] 2025-12-04T09:49:26.2982452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2982495Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2982549Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2982623Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2982854Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2982891Z graph_break [] 2025-12-04T09:49:26.2982964Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2983006Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2983060Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2983132Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2983362Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2983427Z graph_break [] 2025-12-04T09:49:26.2983499Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2983542Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2983596Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2983669Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2983896Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2983933Z graph_break [] 2025-12-04T09:49:26.2984006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2984048Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2984105Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2984177Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2984406Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2984442Z graph_break [] 2025-12-04T09:49:26.2984518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2984558Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2984614Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2984685Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2984933Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2984972Z graph_break [] 2025-12-04T09:49:26.2985045Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2985087Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2985142Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2985213Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2985442Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2985479Z graph_break [] 2025-12-04T09:49:26.2985553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2985594Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2985649Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2985722Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2985951Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2985989Z graph_break [] 2025-12-04T09:49:26.2986062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2986102Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2986158Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2986231Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2986460Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2986496Z graph_break [] 2025-12-04T09:49:26.2986571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2986625Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2986680Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2986802Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2987033Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2987069Z graph_break [] 2025-12-04T09:49:26.2987143Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2987184Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2987239Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2987310Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2987542Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2987580Z graph_break [] 2025-12-04T09:49:26.2987654Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2987695Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2987750Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2987822Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2988052Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2988088Z graph_break [] 2025-12-04T09:49:26.2988174Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.2988221Z Traceback (most recent call last): 2025-12-04T09:49:26.2988365Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.2988407Z return value(self) 2025-12-04T09:49:26.2988551Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.2988603Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.2988741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.2988803Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.2988964Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.2989037Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.2989089Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.2989091Z 2025-12-04T09:49:26.2989143Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.2989262Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.2989362Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.2989365Z 2025-12-04T09:49:26.2989438Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.2989590Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.2989593Z 2025-12-04T09:49:26.2989680Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.2989754Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2989796Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2989852Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2990082Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2990173Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2990222Z graph_break [] 2025-12-04T09:49:26.2990296Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2990337Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2990393Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2990463Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2990692Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2990729Z graph_break [] 2025-12-04T09:49:26.2990801Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2990845Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2990900Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2990972Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2991201Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2991239Z graph_break [] 2025-12-04T09:49:26.2991311Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2991353Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2991407Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2991479Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2991724Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2991762Z graph_break [] 2025-12-04T09:49:26.2991834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2991877Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2991931Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2992004Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2992231Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2992268Z graph_break [] 2025-12-04T09:49:26.2992340Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2992382Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2992436Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2992509Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2992740Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2992778Z graph_break [] 2025-12-04T09:49:26.2992851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2992892Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2992945Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2993017Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2993245Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2993283Z graph_break [] 2025-12-04T09:49:26.2993357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2993411Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2993479Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2993551Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2993779Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2993816Z graph_break [] 2025-12-04T09:49:26.2993888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2993929Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2993984Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2994055Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2994285Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2994324Z graph_break [] 2025-12-04T09:49:26.2994398Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2994438Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2994493Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2994567Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2994801Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2994837Z graph_break [] 2025-12-04T09:49:26.2994913Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2994975Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2995031Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2995105Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2995334Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2995371Z graph_break [] 2025-12-04T09:49:26.2995444Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2995485Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2995541Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2995613Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2995845Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2995882Z graph_break [] 2025-12-04T09:49:26.2995956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2995998Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2996054Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2996125Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2996353Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2996389Z graph_break [] 2025-12-04T09:49:26.2996463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2996503Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2996558Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2996631Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2996945Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2996999Z graph_break [] 2025-12-04T09:49:26.2997072Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2997113Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2997168Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2997238Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2997467Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2997503Z graph_break [] 2025-12-04T09:49:26.2997577Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2997619Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2997673Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2997745Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2997974Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2998011Z graph_break [] 2025-12-04T09:49:26.2998084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2998125Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2998180Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2998252Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2998502Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2998542Z graph_break [] 2025-12-04T09:49:26.2998616Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2998658Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2998712Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2998784Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2999013Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2999050Z graph_break [] 2025-12-04T09:49:26.2999122Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2999164Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2999220Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2999294Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.2999525Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.2999562Z graph_break [] 2025-12-04T09:49:26.2999635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.2999679Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.2999733Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.2999806Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3000034Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3000073Z graph_break [] 2025-12-04T09:49:26.3000159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3000213Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3000267Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3000340Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3000568Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3000607Z graph_break [] 2025-12-04T09:49:26.3000681Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3000724Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3000778Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3000853Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3001079Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3001122Z graph_break [] 2025-12-04T09:49:26.3001196Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3001239Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3001293Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3001366Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3001593Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3001630Z graph_break [] 2025-12-04T09:49:26.3001721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3001763Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3001819Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3001895Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3002127Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3002164Z graph_break [] 2025-12-04T09:49:26.3002240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3002281Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3002338Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3002409Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3002640Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3002676Z graph_break [] 2025-12-04T09:49:26.3002756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3002797Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3002854Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3002926Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3003156Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3003192Z graph_break [] 2025-12-04T09:49:26.3003266Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3003308Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3003366Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3003449Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3003681Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3003727Z graph_break [] 2025-12-04T09:49:26.3003801Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3003842Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3003896Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3003967Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3004200Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3004237Z graph_break [] 2025-12-04T09:49:26.3004310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3004352Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3004410Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3004482Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3004710Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3004748Z graph_break [] 2025-12-04T09:49:26.3004824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3004867Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3004922Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3004994Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3005242Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3005282Z graph_break [] 2025-12-04T09:49:26.3005358Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3005398Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3005454Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3005525Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3005752Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3005792Z graph_break [] 2025-12-04T09:49:26.3005864Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3005909Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3005963Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3006035Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3006263Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3006303Z graph_break [] 2025-12-04T09:49:26.3006376Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3006418Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3006472Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3006546Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3006823Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3006893Z graph_break [] 2025-12-04T09:49:26.3006966Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3007025Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3007080Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3007153Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3007380Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3007416Z graph_break [] 2025-12-04T09:49:26.3007488Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3007530Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3007584Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3007658Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3007886Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3007926Z graph_break [] 2025-12-04T09:49:26.3007998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3008040Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3008095Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3008170Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3008397Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3008436Z graph_break [] 2025-12-04T09:49:26.3008540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3008586Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3008640Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3008717Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3008946Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3008984Z graph_break [] 2025-12-04T09:49:26.3009057Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3009100Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3009155Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3009227Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3009457Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3009496Z graph_break [] 2025-12-04T09:49:26.3009570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3009611Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3009666Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3009738Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3009967Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3010005Z graph_break [] 2025-12-04T09:49:26.3010078Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3010121Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3010177Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3010265Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3010509Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3010545Z graph_break [] 2025-12-04T09:49:26.3010618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3010659Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3010715Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3010786Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3011022Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3011060Z graph_break [] 2025-12-04T09:49:26.3011134Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3011175Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3011231Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3011302Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3011531Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3011567Z graph_break [] 2025-12-04T09:49:26.3011641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3011681Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3011736Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3011830Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3012062Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3012101Z graph_break [] 2025-12-04T09:49:26.3012175Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3012217Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3012272Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3012345Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3012573Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3012611Z graph_break [] 2025-12-04T09:49:26.3012687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3012730Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3012790Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3012863Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3013092Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3013131Z graph_break [] 2025-12-04T09:49:26.3013203Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3013246Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3013302Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3013375Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3013604Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3013666Z graph_break [] 2025-12-04T09:49:26.3013738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3013782Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3013837Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3013912Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3014138Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3014176Z graph_break [] 2025-12-04T09:49:26.3014261Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.3014308Z Traceback (most recent call last): 2025-12-04T09:49:26.3014429Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.3014469Z return value(self) 2025-12-04T09:49:26.3014618Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.3014672Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.3014812Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.3014876Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.3015040Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.3015116Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.3015169Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3015171Z 2025-12-04T09:49:26.3015250Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3015369Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3015472Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3015475Z 2025-12-04T09:49:26.3015551Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3015704Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3015706Z 2025-12-04T09:49:26.3015794Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3015871Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3015913Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3015970Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3016202Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3016277Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3016316Z graph_break [] 2025-12-04T09:49:26.3016389Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3016435Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3016496Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3016571Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3016883Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3016922Z graph_break [] 2025-12-04T09:49:26.3016998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3017042Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3017123Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3017210Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3017439Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3017477Z graph_break [] 2025-12-04T09:49:26.3017550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3017595Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3017649Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3017721Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3017952Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3017990Z graph_break [] 2025-12-04T09:49:26.3018065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3018108Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3018163Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3018236Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3018464Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3018503Z graph_break [] 2025-12-04T09:49:26.3018575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3018619Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3018724Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3018800Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3019031Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3019071Z graph_break [] 2025-12-04T09:49:26.3019145Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3019190Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3019246Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3019322Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3019551Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3019592Z graph_break [] 2025-12-04T09:49:26.3019667Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3019713Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3019770Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3019843Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3020075Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3020112Z graph_break [] 2025-12-04T09:49:26.3020187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3020229Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3020287Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3020360Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3020594Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3020655Z graph_break [] 2025-12-04T09:49:26.3020729Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3020770Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3020825Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3020896Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3021127Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3021163Z graph_break [] 2025-12-04T09:49:26.3021236Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3021280Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3021337Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3021407Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3021638Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3021674Z graph_break [] 2025-12-04T09:49:26.3021748Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3021788Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3021844Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3021914Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3022179Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3022217Z graph_break [] 2025-12-04T09:49:26.3022293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3022336Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3022391Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3022463Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3022694Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3022732Z graph_break [] 2025-12-04T09:49:26.3022807Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3022848Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3022904Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3022977Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3023212Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3023251Z graph_break [] 2025-12-04T09:49:26.3023326Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3023367Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3023423Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3023495Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3023725Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3023763Z graph_break [] 2025-12-04T09:49:26.3023838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3023893Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3023957Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3024029Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3024256Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3024294Z graph_break [] 2025-12-04T09:49:26.3024368Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3024410Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3024465Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3024536Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3024766Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3024805Z graph_break [] 2025-12-04T09:49:26.3024878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3024920Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3024973Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3025047Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3025277Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3025315Z graph_break [] 2025-12-04T09:49:26.3025387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3025458Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3025514Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3025588Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3025817Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3025854Z graph_break [] 2025-12-04T09:49:26.3025927Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3025970Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3026024Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3026097Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3026326Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3026367Z graph_break [] 2025-12-04T09:49:26.3026441Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3026484Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3026538Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3026611Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3026886Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3026924Z graph_break [] 2025-12-04T09:49:26.3026997Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3027039Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3027093Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3027168Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3027420Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3027470Z graph_break [] 2025-12-04T09:49:26.3027544Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3027585Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3027641Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3027712Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3027940Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3027976Z graph_break [] 2025-12-04T09:49:26.3028051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3028093Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3028148Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3028221Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3028450Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3028486Z graph_break [] 2025-12-04T09:49:26.3028560Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3028601Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3028656Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3028728Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3028985Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3029024Z graph_break [] 2025-12-04T09:49:26.3029100Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3029141Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3029197Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3029269Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3029502Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3029538Z graph_break [] 2025-12-04T09:49:26.3029612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3029652Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3029709Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3029782Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3030014Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3030051Z graph_break [] 2025-12-04T09:49:26.3030125Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3030166Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3030222Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3030293Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3030525Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3030562Z graph_break [] 2025-12-04T09:49:26.3030646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3030700Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3030756Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3030827Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3031054Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3031090Z graph_break [] 2025-12-04T09:49:26.3031164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3031205Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3031260Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3031334Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3031561Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3031601Z graph_break [] 2025-12-04T09:49:26.3031674Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3031716Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3031770Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3031841Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3032069Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3032106Z graph_break [] 2025-12-04T09:49:26.3032197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3032239Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3032295Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3032368Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3032597Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3032634Z graph_break [] 2025-12-04T09:49:26.3032708Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3032749Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3032804Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3032877Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3033108Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3033145Z graph_break [] 2025-12-04T09:49:26.3033218Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3033260Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3033314Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3033387Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3033615Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3033652Z graph_break [] 2025-12-04T09:49:26.3033724Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3033766Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3033822Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3033894Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3034133Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3034188Z graph_break [] 2025-12-04T09:49:26.3034260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3034302Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3034356Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3034429Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3034657Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3034696Z graph_break [] 2025-12-04T09:49:26.3034769Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3034812Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3034867Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3034941Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3035171Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3035207Z graph_break [] 2025-12-04T09:49:26.3035281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3035322Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3035380Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3035453Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3035705Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3035745Z graph_break [] 2025-12-04T09:49:26.3035818Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3035859Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3035915Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3035986Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3036214Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3036251Z graph_break [] 2025-12-04T09:49:26.3036324Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3036366Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3036422Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3036493Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3036724Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3036803Z graph_break [] 2025-12-04T09:49:26.3036877Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3036919Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3036974Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3037045Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3037277Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3037336Z graph_break [] 2025-12-04T09:49:26.3037410Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3037474Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3037530Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3037601Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3037829Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3037866Z graph_break [] 2025-12-04T09:49:26.3037941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3037983Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3038038Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3038110Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3038341Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3038379Z graph_break [] 2025-12-04T09:49:26.3038452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3038493Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3038547Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3038618Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3038845Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3038881Z graph_break [] 2025-12-04T09:49:26.3038978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3039022Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3039078Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3039150Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3039380Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3039417Z graph_break [] 2025-12-04T09:49:26.3039490Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3039532Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3039586Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3039658Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3039886Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3039926Z graph_break [] 2025-12-04T09:49:26.3039998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3040041Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3040095Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3040168Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3040395Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3040432Z graph_break [] 2025-12-04T09:49:26.3040504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3040548Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3040602Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3040685Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3040922Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3040959Z graph_break [] 2025-12-04T09:49:26.3041045Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.3041091Z Traceback (most recent call last): 2025-12-04T09:49:26.3041211Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.3041251Z return value(self) 2025-12-04T09:49:26.3041396Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.3041449Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.3041588Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.3041651Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.3041810Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.3041883Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.3041936Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3041938Z 2025-12-04T09:49:26.3041991Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3042108Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3042206Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3042208Z 2025-12-04T09:49:26.3042303Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3042455Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3042458Z 2025-12-04T09:49:26.3042545Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3042619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3042661Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3042717Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3042947Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3043020Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3043056Z graph_break [] 2025-12-04T09:49:26.3043131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3043174Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3043229Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3043302Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3043529Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3043567Z graph_break [] 2025-12-04T09:49:26.3043643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3045093Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3045151Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3045225Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3045459Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3045515Z graph_break [] 2025-12-04T09:49:26.3045603Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3045646Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3045701Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3045773Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3046001Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3046039Z graph_break [] 2025-12-04T09:49:26.3046111Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3046154Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3046209Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3046285Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3046513Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3046554Z graph_break [] 2025-12-04T09:49:26.3046628Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3046670Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3046724Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3046869Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3047100Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3047177Z graph_break [] 2025-12-04T09:49:26.3047252Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3047294Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3047351Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3047424Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3047654Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3047690Z graph_break [] 2025-12-04T09:49:26.3047764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3047804Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3047860Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3047931Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3048164Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3048202Z graph_break [] 2025-12-04T09:49:26.3048276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3048317Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3048372Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3048442Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3048672Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3048708Z graph_break [] 2025-12-04T09:49:26.3048784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3048825Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3048907Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3048996Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3049226Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3049262Z graph_break [] 2025-12-04T09:49:26.3049336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3049377Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3049432Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3049504Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3049733Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3049771Z graph_break [] 2025-12-04T09:49:26.3049844Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3049886Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3049941Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3050012Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3050240Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3050276Z graph_break [] 2025-12-04T09:49:26.3050350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3050390Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3050469Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3050541Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3050773Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3050811Z graph_break [] 2025-12-04T09:49:26.3050885Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3050926Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3050981Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3051052Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3051281Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3051320Z graph_break [] 2025-12-04T09:49:26.3051392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3051435Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3051491Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3051564Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3051793Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3051830Z graph_break [] 2025-12-04T09:49:26.3051903Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3051944Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3052001Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3052073Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3052302Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3052361Z graph_break [] 2025-12-04T09:49:26.3052434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3052476Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3052531Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3052604Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3052835Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3052872Z graph_break [] 2025-12-04T09:49:26.3052945Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3052989Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3053045Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3053118Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3053352Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3053390Z graph_break [] 2025-12-04T09:49:26.3053464Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3053506Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3053560Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3053633Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3053883Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3053923Z graph_break [] 2025-12-04T09:49:26.3053996Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3054040Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3054096Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3054169Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3054395Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3054432Z graph_break [] 2025-12-04T09:49:26.3054505Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3054547Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3054602Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3054675Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3054904Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3054942Z graph_break [] 2025-12-04T09:49:26.3055016Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3055057Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3055112Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3055183Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3055410Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3055446Z graph_break [] 2025-12-04T09:49:26.3055521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3055572Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3055627Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3055709Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3055939Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3055975Z graph_break [] 2025-12-04T09:49:26.3056049Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3056090Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3056148Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3056219Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3056452Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3056490Z graph_break [] 2025-12-04T09:49:26.3056563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3056604Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3056659Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3056731Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3057012Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3057048Z graph_break [] 2025-12-04T09:49:26.3057121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3057198Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3057254Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3057328Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3057557Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3057593Z graph_break [] 2025-12-04T09:49:26.3057668Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3057709Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3057765Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3057836Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3058066Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3058103Z graph_break [] 2025-12-04T09:49:26.3058176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3058218Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3058274Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3058344Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3058573Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3058610Z graph_break [] 2025-12-04T09:49:26.3058681Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3058723Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3058778Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3058851Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3059101Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3059154Z graph_break [] 2025-12-04T09:49:26.3059227Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3059268Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3059323Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3059394Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3059623Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3059660Z graph_break [] 2025-12-04T09:49:26.3059734Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3059777Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3059831Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3059907Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3060136Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3060173Z graph_break [] 2025-12-04T09:49:26.3060246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3060287Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3060342Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3060416Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3060679Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3060718Z graph_break [] 2025-12-04T09:49:26.3060791Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3060833Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3060888Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3060962Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3061189Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3061226Z graph_break [] 2025-12-04T09:49:26.3061298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3061340Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3061395Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3061468Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3061697Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3061734Z graph_break [] 2025-12-04T09:49:26.3061806Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3061848Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3061903Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3061975Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3062206Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3062244Z graph_break [] 2025-12-04T09:49:26.3062328Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3062382Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3062437Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3062509Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3062736Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3062774Z graph_break [] 2025-12-04T09:49:26.3062847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3062888Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3062944Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3063017Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3063246Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3063284Z graph_break [] 2025-12-04T09:49:26.3063358Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3063399Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3063454Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3063525Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3063755Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3063791Z graph_break [] 2025-12-04T09:49:26.3063885Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3063927Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3063984Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3064056Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3064285Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3064321Z graph_break [] 2025-12-04T09:49:26.3064395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3064435Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3064492Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3064562Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3064790Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3064827Z graph_break [] 2025-12-04T09:49:26.3064902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3064943Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3064998Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3065070Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3065297Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3065334Z graph_break [] 2025-12-04T09:49:26.3065407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3065447Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3065505Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3065577Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3065818Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3065868Z graph_break [] 2025-12-04T09:49:26.3065941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3065982Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3066037Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3066108Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3066338Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3066377Z graph_break [] 2025-12-04T09:49:26.3066451Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3066496Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3066552Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3066624Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3066903Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3066941Z graph_break [] 2025-12-04T09:49:26.3067015Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3067057Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3067112Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3067183Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3067441Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3067483Z graph_break [] 2025-12-04T09:49:26.3067555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3067597Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3067652Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3067726Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3067953Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3067994Z graph_break [] 2025-12-04T09:49:26.3068070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3068115Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3068171Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3068243Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3068475Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3068512Z graph_break [] 2025-12-04T09:49:26.3068585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3068630Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3068684Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3068758Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3068987Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3069040Z graph_break [] 2025-12-04T09:49:26.3069112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3069171Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3069227Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3069303Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3069529Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3069566Z graph_break [] 2025-12-04T09:49:26.3069651Z ________________ FreezingCpuTests.test_conv_with_as_strided_cpu ________________ 2025-12-04T09:49:26.3069701Z Traceback (most recent call last): 2025-12-04T09:49:26.3069825Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:49:26.3069869Z return value(self) 2025-12-04T09:49:26.3070014Z File "/var/lib/jenkins/pytorch/test/inductor/test_inductor_freezing.py", line 789, in test_conv_with_as_strided 2025-12-04T09:49:26.3070068Z self.assertEqual(foo(mod, x), mod_eager) 2025-12-04T09:49:26.3070209Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:49:26.3070272Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:49:26.3070435Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:49:26.3070508Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:49:26.3070563Z AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3070566Z 2025-12-04T09:49:26.3070617Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3070759Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3070861Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3070865Z 2025-12-04T09:49:26.3070940Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3071096Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3071099Z 2025-12-04T09:49:26.3071185Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3071260Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3071302Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3071360Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3071592Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3071670Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3071707Z graph_break [] 2025-12-04T09:49:26.3071781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3071822Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3071877Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3071949Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3072179Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3072215Z graph_break [] 2025-12-04T09:49:26.3072291Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3072331Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3072390Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3072475Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3072718Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3072756Z graph_break [] 2025-12-04T09:49:26.3072830Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3072871Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3072929Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3073001Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3073237Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3073275Z graph_break [] 2025-12-04T09:49:26.3073351Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3073393Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3073450Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3073523Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3073753Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3073791Z graph_break [] 2025-12-04T09:49:26.3073864Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3073907Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3073961Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3074061Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3074290Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3074330Z graph_break [] 2025-12-04T09:49:26.3074404Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3074447Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3074504Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3074576Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3074805Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3074841Z graph_break [] 2025-12-04T09:49:26.3074915Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3074957Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3075012Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3075085Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3075316Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3075354Z graph_break [] 2025-12-04T09:49:26.3075429Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3075471Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3075528Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3075599Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3075830Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3075879Z graph_break [] 2025-12-04T09:49:26.3075967Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3076009Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3076065Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3076138Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3076370Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3076407Z graph_break [] 2025-12-04T09:49:26.3076480Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3076521Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3076578Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3076653Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3076917Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3076956Z graph_break [] 2025-12-04T09:49:26.3077028Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3077073Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3077128Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3077201Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3077430Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3077495Z graph_break [] 2025-12-04T09:49:26.3077568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3077612Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3077667Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3077739Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3077967Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3078005Z graph_break [] 2025-12-04T09:49:26.3078079Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3078120Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3078174Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3078245Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3078474Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3078513Z graph_break [] 2025-12-04T09:49:26.3078587Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3078629Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3078683Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3078757Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3078984Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3079021Z graph_break [] 2025-12-04T09:49:26.3079093Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3079138Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3079210Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3079283Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3079523Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3079562Z graph_break [] 2025-12-04T09:49:26.3079636Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3079678Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3079733Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3079806Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3080036Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3080075Z graph_break [] 2025-12-04T09:49:26.3080148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3080190Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3080244Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3080318Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3080547Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3080583Z graph_break [] 2025-12-04T09:49:26.3080657Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3080699Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3080774Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3080848Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3081077Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3081114Z graph_break [] 2025-12-04T09:49:26.3081187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3081228Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3081283Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3081355Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3081585Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3081621Z graph_break [] 2025-12-04T09:49:26.3081696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3081738Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3081796Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3081867Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3082098Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3082134Z graph_break [] 2025-12-04T09:49:26.3082208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3082249Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3082304Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3082375Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3082605Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3082662Z graph_break [] 2025-12-04T09:49:26.3082737Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3082777Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3082832Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3082903Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3083134Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3083170Z graph_break [] 2025-12-04T09:49:26.3083244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3083286Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3083342Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3083415Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3083647Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3083684Z graph_break [] 2025-12-04T09:49:26.3083757Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3083798Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3083854Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3083926Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3084175Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3084213Z graph_break [] 2025-12-04T09:49:26.3084286Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3084328Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3084382Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3084454Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3084683Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3084720Z graph_break [] 2025-12-04T09:49:26.3084792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3084834Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3084888Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3084963Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3085193Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3085231Z graph_break [] 2025-12-04T09:49:26.3085303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3085345Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3085399Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3085471Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3085700Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3085737Z graph_break [] 2025-12-04T09:49:26.3085811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3085864Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3085919Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3086010Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3086236Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3086274Z graph_break [] 2025-12-04T09:49:26.3086345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3086387Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3086441Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3086514Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3086783Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3086822Z graph_break [] 2025-12-04T09:49:26.3086894Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3086936Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3086990Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3087062Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3087289Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3087326Z graph_break [] 2025-12-04T09:49:26.3087398Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3087440Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3087522Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3087596Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3087829Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3087865Z graph_break [] 2025-12-04T09:49:26.3087939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3087979Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3088034Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3088105Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3088334Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3088371Z graph_break [] 2025-12-04T09:49:26.3088445Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3088487Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3088543Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3088614Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3088842Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3088878Z graph_break [] 2025-12-04T09:49:26.3088951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3088992Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3089047Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3089121Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3089365Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3089415Z graph_break [] 2025-12-04T09:49:26.3089488Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3089528Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3089583Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3089654Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3089884Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3089920Z graph_break [] 2025-12-04T09:49:26.3089994Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3090035Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3090091Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3090163Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3090392Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3090428Z graph_break [] 2025-12-04T09:49:26.3090501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3090541Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3090596Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3090667Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3090917Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3090955Z graph_break [] 2025-12-04T09:49:26.3091030Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3091071Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3091126Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3091197Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3091427Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3091464Z graph_break [] 2025-12-04T09:49:26.3091536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3091577Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3091633Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3091707Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3091936Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3091975Z graph_break [] 2025-12-04T09:49:26.3092048Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3092090Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3092145Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3092218Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3092447Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3092486Z graph_break [] 2025-12-04T09:49:26.3092570Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3092622Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3092676Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3092748Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3092975Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3093012Z graph_break [] 2025-12-04T09:49:26.3093084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3093126Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3093180Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3093253Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3093479Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3093518Z graph_break [] 2025-12-04T09:49:26.3093592Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3093633Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3093688Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3093759Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3093985Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3094024Z graph_break [] 2025-12-04T09:49:26.3094114Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3094156Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3094211Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3094284Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3094511Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3094547Z graph_break [] 2025-12-04T09:49:26.3094620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3094661Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3094715Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3094787Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3095015Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3095054Z graph_break [] 2025-12-04T09:49:26.3095128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3095168Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3095224Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3095295Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3095522Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3095558Z graph_break [] 2025-12-04T09:49:26.3095633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3095673Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3095731Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3095802Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3096043Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3096089Z graph_break [] 2025-12-04T09:49:26.3096163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3096204Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3096260Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3096331Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3096559Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3096597Z graph_break [] 2025-12-04T09:49:26.3096670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:49:26.3096712Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:49:26.3096815Z stats [('calls_captured', 5), ('unique_graphs', 1)] 2025-12-04T09:49:26.3096887Z aot_autograd [('total', 1), ('autograd_cache_bypass', 1), ('ok', 1)] 2025-12-04T09:49:26.3097118Z inductor [('extern_calls', 2), ('mkldnn_conv_weight_pack_matcher_count', 1), ('mkldnn_conv_weight_pack_matcher_nodes', 1), ('pattern_matcher_count', 1), ('pattern_matcher_nodes', 1), ('fxgraph_cache_bypass', 1)] 2025-12-04T09:49:26.3097154Z graph_break [] 2025-12-04T09:49:26.3097394Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_inductor_freezing/inductor.test_inductor_freezing-82f27a0beef4242a.xml - 2025-12-04T09:49:26.3097456Z =========================== short test summary info ============================ 2025-12-04T09:49:26.3097697Z FAILED [4.7162s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3097700Z 2025-12-04T09:49:26.3097753Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3097873Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3097972Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3097975Z 2025-12-04T09:49:26.3098048Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3098200Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3098202Z 2025-12-04T09:49:26.3098289Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3098474Z FAILED [2.4033s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3098476Z 2025-12-04T09:49:26.3098527Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3098644Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3098740Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3098743Z 2025-12-04T09:49:26.3098815Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3098964Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3098966Z 2025-12-04T09:49:26.3099051Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3099231Z FAILED [2.1367s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3099233Z 2025-12-04T09:49:26.3099286Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3099424Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3099532Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3099534Z 2025-12-04T09:49:26.3099606Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3099754Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3099756Z 2025-12-04T09:49:26.3099841Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3100021Z FAILED [2.3047s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3100023Z 2025-12-04T09:49:26.3100074Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3100190Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3100286Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3100289Z 2025-12-04T09:49:26.3100360Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3100510Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3100512Z 2025-12-04T09:49:26.3100595Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3100774Z FAILED [2.3065s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3100777Z 2025-12-04T09:49:26.3100827Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3100941Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3101054Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3101058Z 2025-12-04T09:49:26.3101129Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3101279Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3101282Z 2025-12-04T09:49:26.3101365Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3101544Z FAILED [1.8926s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3101546Z 2025-12-04T09:49:26.3101596Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3101710Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3101806Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3101808Z 2025-12-04T09:49:26.3101880Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3102028Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3102031Z 2025-12-04T09:49:26.3102115Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3102294Z FAILED [1.7583s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3102297Z 2025-12-04T09:49:26.3102347Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3102461Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3102555Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3102557Z 2025-12-04T09:49:26.3102630Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3102792Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3102806Z 2025-12-04T09:49:26.3102891Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3103068Z FAILED [1.9130s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3103070Z 2025-12-04T09:49:26.3103121Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3103234Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3103329Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3103331Z 2025-12-04T09:49:26.3103401Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3103552Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3103555Z 2025-12-04T09:49:26.3103638Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3103819Z FAILED [1.3642s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3103821Z 2025-12-04T09:49:26.3103872Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3103985Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3104083Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3104085Z 2025-12-04T09:49:26.3104155Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3104326Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3104329Z 2025-12-04T09:49:26.3104413Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3104592Z FAILED [2.4604s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3104594Z 2025-12-04T09:49:26.3104646Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3104761Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3104855Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3104857Z 2025-12-04T09:49:26.3104929Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3105078Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3105080Z 2025-12-04T09:49:26.3105166Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3105345Z FAILED [2.3937s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3105349Z 2025-12-04T09:49:26.3105399Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3105518Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3105613Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3105615Z 2025-12-04T09:49:26.3105686Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3105835Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3105838Z 2025-12-04T09:49:26.3105922Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3106104Z FAILED [0.8786s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3106127Z 2025-12-04T09:49:26.3106179Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3106291Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3106386Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3106388Z 2025-12-04T09:49:26.3106459Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3106728Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3106731Z 2025-12-04T09:49:26.3106854Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3107036Z FAILED [2.0473s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3107039Z 2025-12-04T09:49:26.3107091Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3107205Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3107301Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3107303Z 2025-12-04T09:49:26.3107374Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3107522Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3107523Z 2025-12-04T09:49:26.3107607Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3107785Z FAILED [1.8847s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3107819Z 2025-12-04T09:49:26.3107871Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3107986Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3108081Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3108083Z 2025-12-04T09:49:26.3108156Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3108303Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3108306Z 2025-12-04T09:49:26.3108390Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3108569Z FAILED [2.4815s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3108571Z 2025-12-04T09:49:26.3108622Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3108737Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3108832Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3108835Z 2025-12-04T09:49:26.3108907Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3109054Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3109056Z 2025-12-04T09:49:26.3109140Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3109317Z FAILED [1.3913s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3109319Z 2025-12-04T09:49:26.3109371Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3109486Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3109594Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3109609Z 2025-12-04T09:49:26.3109681Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3109831Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3109832Z 2025-12-04T09:49:26.3109917Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3110096Z FAILED [1.6226s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3110097Z 2025-12-04T09:49:26.3110148Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3110262Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3110359Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3110362Z 2025-12-04T09:49:26.3110433Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3110583Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3110585Z 2025-12-04T09:49:26.3110668Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3110847Z FAILED [2.6011s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3110849Z 2025-12-04T09:49:26.3110899Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3111014Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3111108Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3111135Z 2025-12-04T09:49:26.3111207Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3111357Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3111361Z 2025-12-04T09:49:26.3111444Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3111623Z FAILED [0.8914s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3111625Z 2025-12-04T09:49:26.3111675Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3111789Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3111882Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3111884Z 2025-12-04T09:49:26.3111957Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3112103Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3112107Z 2025-12-04T09:49:26.3112191Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3112369Z FAILED [2.3420s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3112371Z 2025-12-04T09:49:26.3112422Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3112538Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3112634Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3112636Z 2025-12-04T09:49:26.3112707Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3112856Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3112869Z 2025-12-04T09:49:26.3112954Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3113142Z FAILED [0.9698s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3113144Z 2025-12-04T09:49:26.3113195Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3113308Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3113403Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3113405Z 2025-12-04T09:49:26.3113475Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3113623Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3113627Z 2025-12-04T09:49:26.3113710Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3113890Z FAILED [1.6286s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3113892Z 2025-12-04T09:49:26.3113943Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3114058Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3114151Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3114155Z 2025-12-04T09:49:26.3114225Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3114374Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3114376Z 2025-12-04T09:49:26.3114479Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3114661Z FAILED [1.8100s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3114664Z 2025-12-04T09:49:26.3114714Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3114828Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3114921Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3114923Z 2025-12-04T09:49:26.3114995Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3115142Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3115144Z 2025-12-04T09:49:26.3115229Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3115408Z FAILED [1.7601s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3115411Z 2025-12-04T09:49:26.3115463Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3115576Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3115672Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3115674Z 2025-12-04T09:49:26.3115746Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3115892Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3115894Z 2025-12-04T09:49:26.3115978Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3116157Z FAILED [2.1848s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3116168Z 2025-12-04T09:49:26.3116220Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3116333Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3116442Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3116444Z 2025-12-04T09:49:26.3116515Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3116663Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3116665Z 2025-12-04T09:49:26.3116790Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3116971Z FAILED [1.3902s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3116973Z 2025-12-04T09:49:26.3117025Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3117141Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3117236Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3117239Z 2025-12-04T09:49:26.3117309Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3117457Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3117459Z 2025-12-04T09:49:26.3117542Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3117721Z FAILED [1.6944s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3117723Z 2025-12-04T09:49:26.3117773Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3117917Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3118012Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3118016Z 2025-12-04T09:49:26.3118087Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3118234Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3118236Z 2025-12-04T09:49:26.3118321Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3118499Z FAILED [1.2763s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3118501Z 2025-12-04T09:49:26.3118552Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3118667Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3118762Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3118766Z 2025-12-04T09:49:26.3118838Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3118986Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3118988Z 2025-12-04T09:49:26.3119072Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3119249Z FAILED [2.2083s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3119251Z 2025-12-04T09:49:26.3119302Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3119415Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3119513Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3119527Z 2025-12-04T09:49:26.3119598Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3119759Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3119761Z 2025-12-04T09:49:26.3119844Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3120022Z FAILED [2.2002s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3120024Z 2025-12-04T09:49:26.3120075Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3120190Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3120284Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3120286Z 2025-12-04T09:49:26.3120357Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3120507Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3120510Z 2025-12-04T09:49:26.3120593Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3120773Z FAILED [2.0791s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3120774Z 2025-12-04T09:49:26.3120825Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3120939Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3121033Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3121035Z 2025-12-04T09:49:26.3121106Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3121273Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3121277Z 2025-12-04T09:49:26.3121362Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3121541Z FAILED [1.2334s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3121544Z 2025-12-04T09:49:26.3121595Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3121710Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3121803Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3121805Z 2025-12-04T09:49:26.3121877Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3122026Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3122029Z 2025-12-04T09:49:26.3122113Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3122291Z FAILED [1.9387s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3122293Z 2025-12-04T09:49:26.3122344Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3122457Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3122552Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3122554Z 2025-12-04T09:49:26.3122624Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3122774Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3122775Z 2025-12-04T09:49:26.3122860Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3123048Z FAILED [1.6752s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3123060Z 2025-12-04T09:49:26.3123111Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3123225Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3123319Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3123321Z 2025-12-04T09:49:26.3123391Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3123539Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3123540Z 2025-12-04T09:49:26.3123625Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3123805Z FAILED [1.9322s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3123809Z 2025-12-04T09:49:26.3123859Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3123976Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3124069Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3124071Z 2025-12-04T09:49:26.3124145Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3124293Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3124295Z 2025-12-04T09:49:26.3124380Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3124589Z FAILED [1.2119s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3124592Z 2025-12-04T09:49:26.3124643Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3124759Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3124854Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3124856Z 2025-12-04T09:49:26.3124929Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3125076Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3125078Z 2025-12-04T09:49:26.3125162Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3125342Z FAILED [2.0943s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3125344Z 2025-12-04T09:49:26.3125398Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3125511Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3125609Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3125611Z 2025-12-04T09:49:26.3125681Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3125829Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3125831Z 2025-12-04T09:49:26.3125916Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3126096Z FAILED [1.4846s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3126098Z 2025-12-04T09:49:26.3126154Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3126266Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3126386Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3126388Z 2025-12-04T09:49:26.3126458Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3126606Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3126608Z 2025-12-04T09:49:26.3126692Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3126925Z FAILED [2.2102s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3126927Z 2025-12-04T09:49:26.3126978Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3127096Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3127190Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3127193Z 2025-12-04T09:49:26.3127267Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3127415Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3127421Z 2025-12-04T09:49:26.3127504Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3127686Z FAILED [2.1972s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3127688Z 2025-12-04T09:49:26.3127738Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3127852Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3127979Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3127982Z 2025-12-04T09:49:26.3128056Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3128205Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3128207Z 2025-12-04T09:49:26.3128293Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3128471Z FAILED [2.1349s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3128473Z 2025-12-04T09:49:26.3128524Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3128637Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3128736Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3128737Z 2025-12-04T09:49:26.3128809Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3128960Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3128963Z 2025-12-04T09:49:26.3129049Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3129229Z FAILED [1.5422s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3129231Z 2025-12-04T09:49:26.3129282Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3129396Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3129490Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3129492Z 2025-12-04T09:49:26.3129564Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3129730Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3129748Z 2025-12-04T09:49:26.3129832Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3130012Z FAILED [2.1583s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3130014Z 2025-12-04T09:49:26.3130063Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3130178Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3130272Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3130274Z 2025-12-04T09:49:26.3130346Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3130496Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3130499Z 2025-12-04T09:49:26.3130583Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3130766Z FAILED [0.9261s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3130768Z 2025-12-04T09:49:26.3130819Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3130936Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3131029Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3131031Z 2025-12-04T09:49:26.3131103Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3131272Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3131274Z 2025-12-04T09:49:26.3131361Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3131542Z FAILED [1.9690s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3131544Z 2025-12-04T09:49:26.3131595Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3131709Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3131804Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3131806Z 2025-12-04T09:49:26.3131879Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3132026Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3132028Z 2025-12-04T09:49:26.3132114Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3132293Z FAILED [1.9531s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3132296Z 2025-12-04T09:49:26.3132347Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3132462Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3132558Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3132560Z 2025-12-04T09:49:26.3132632Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3132782Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3132784Z 2025-12-04T09:49:26.3132868Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3133053Z FAILED [1.7575s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3133074Z 2025-12-04T09:49:26.3133124Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3133239Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3133333Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3133336Z 2025-12-04T09:49:26.3133406Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3133556Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3133559Z 2025-12-04T09:49:26.3133642Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3133823Z FAILED [2.2597s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3133825Z 2025-12-04T09:49:26.3133875Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3133990Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3134083Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3134085Z 2025-12-04T09:49:26.3134156Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3134304Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3134306Z 2025-12-04T09:49:26.3134391Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3134568Z FAILED [0.6902s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3134588Z 2025-12-04T09:49:26.3134640Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3134756Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3134855Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3134856Z 2025-12-04T09:49:26.3134929Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3135077Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3135079Z 2025-12-04T09:49:26.3135163Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3135341Z FAILED [0.9460s] inductor/test_inductor_freezing.py::FreezingCpuTests::test_conv_with_as_strided_cpu - AssertionError: Tensor-likes are not close! 2025-12-04T09:49:26.3135343Z 2025-12-04T09:49:26.3135395Z Mismatched elements: 1574636 / 1769472 (89.0%) 2025-12-04T09:49:26.3135511Z Greatest absolute difference: 3.915588855743408 at index (0, 166, 0, 0, 8, 2) (up to 1e-05 allowed) 2025-12-04T09:49:26.3135608Z Greatest relative difference: inf at index (0, 40, 0, 0, 7, 0) (up to 1.3e-06 allowed) 2025-12-04T09:49:26.3135611Z 2025-12-04T09:49:26.3135682Z To execute this test, run the following from the base repo dir: 2025-12-04T09:49:26.3135832Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_inductor_freezing.py FreezingCpuTests.test_conv_with_as_strided_cpu 2025-12-04T09:49:26.3135833Z 2025-12-04T09:49:26.3135918Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:49:26.3135989Z ================== 50 failed, 50 skipped in 93.48s (0:01:33) =================== 2025-12-04T09:49:26.3135991Z 2025-12-04T09:49:26.3136172Z FINISHED PRINTING LOG FILE of inductor/test_inductor_freezing 1/1 (test/test-reports/inductor.test_inductor_freezing_1.1_b201467f2f95967c_.log) 2025-12-04T09:49:26.3136177Z 2025-12-04T09:49:26.3136299Z Finished inductor/test_inductor_freezing 1/1 ... [2025-12-04 09:49:26.212490][2225590.669803344], took 1.68min 2025-12-04T09:49:26.3136543Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:49:26.3136675Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:49:26.3136815Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-12-04T09:49:26.3136865Z Uploading artifacts took 0.00 seconds 2025-12-04T09:49:26.3136918Z inductor/test_inductor_freezing 1/1 failed! 2025-12-04T09:49:26.3137021Z Running inductor/test_flex_attention 2/4 ... [2025-12-04 09:49:26.224331][2225590.681649697] 2025-12-04T09:49:26.3137073Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:49:26.3137446Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_flex_attention.py', '--shard-id=2', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:49:26.224574] 2025-12-04T09:50:52.3602372Z 2025-12-04T09:50:52.3603408Z inductor/test_flex_attention 2/4 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_flex_attention_2.4_4650ae2d192e8477_.log 2025-12-04T09:50:52.3635278Z Running 100 items in this shard: test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_flex_attention_stride_ordering_mode_paged_attention_permute_order3_shape1_cuda, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16, test/inductor/test_flex_attention.py::TestFlexAttentionCUDA::test_return_aux_deprecation_warnings_cuda_float16 2025-12-04T09:50:52.3652997Z 2025-12-04T09:50:52.3653129Z Finished inductor/test_flex_attention 2/4 ... [2025-12-04 09:50:52.359768][2225676.817089142], took 1.44min 2025-12-04T09:50:52.3653534Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:50:52.3712643Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:50:52.3714514Z Running inductor/test_cpu_cpp_wrapper 1/1 ... [2025-12-04 09:50:52.371287][2225676.828611509] 2025-12-04T09:50:52.3714731Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:50:52.3716038Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_cpu_cpp_wrapper.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:52.371483] 2025-12-04T09:50:59.7398907Z 2025-12-04T09:50:59.7399942Z inductor/test_cpu_cpp_wrapper 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_cpu_cpp_wrapper_1.1_180a87ceb4fb0288_.log 2025-12-04T09:50:59.7400372Z 2025-12-04T09:50:59.7400559Z Finished inductor/test_cpu_cpp_wrapper 1/1 ... [2025-12-04 09:50:59.739578][2225684.196900649], took 0.12min 2025-12-04T09:50:59.7402153Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:50:59.7504564Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:50:59.7506534Z Running inductor/test_compile_subprocess 3/3 ... [2025-12-04 09:50:59.750556][2225684.207879964] 2025-12-04T09:50:59.7506936Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:50:59.7508809Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_compile_subprocess.py', '--shard-id=3', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:50:59.750757] 2025-12-04T09:54:35.0267964Z 2025-12-04T09:54:35.0269051Z PRINTING LOG FILE of inductor/test_compile_subprocess 3/3 (test/test-reports/inductor.test_compile_subprocess_3.3_441e1e3480e0b5ea_.log) 2025-12-04T09:54:35.0269886Z Test results will be stored in test-reports/python-pytest/inductor.test_compile_subprocess/inductor.test_compile_subprocess-90b2f4794b3bee89.xml 2025-12-04T09:54:35.0270440Z ============================= test session starts ============================== 2025-12-04T09:54:35.0270859Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T09:54:35.0271217Z cachedir: .pytest_cache 2025-12-04T09:54:35.0271688Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T09:54:35.0272141Z rootdir: /var/lib/jenkins/pytorch 2025-12-04T09:54:35.0272364Z configfile: pytest.ini 2025-12-04T09:54:35.0272790Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T09:54:35.0273232Z collecting ... collected 897 items 2025-12-04T09:54:35.0273493Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T09:54:35.0303264Z Running 150 items in this shard: test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_progressive, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda, test/inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.0320365Z 2025-12-04T09:54:35.0321281Z inductor/test_compile_subprocess.py::GPUTests::test_progressive E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] Exception No valid triton configs. OutOfMemoryError: out of resource: triton_mm Required: 65536 Hardware limit:65536 Reducing block sizes or `num_stages` may help. for benchmark choice TritonTemplateCaller(/tmp/tmp8d0mvaed/ps/cpswvyhocuwz36r4wtbcavgybdba2dgixgl3jdou4yu4rvsjfv2o.py, ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=256, BLOCK_M=64, BLOCK_N=64, EVEN_K=True, GROUP_M=4, USE_FAST_ACCUM=False, kpack=2, matrix_instr_nonkdim=16, waves_per_eu=0, num_stages=2, num_warps=8) 2025-12-04T09:54:35.0322291Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] Traceback (most recent call last): 2025-12-04T09:54:35.0322696Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/concurrent/futures/thread.py", line 58, in run 2025-12-04T09:54:35.0323112Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] result = self.fn(*self.args, **self.kwargs) 2025-12-04T09:54:35.0323599Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 3255, in precompile_with_captured_stdout 2025-12-04T09:54:35.0324056Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] choice.precompile() 2025-12-04T09:54:35.0324477Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2289, in precompile 2025-12-04T09:54:35.0324902Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] self.bmreq.precompile() 2025-12-04T09:54:35.0325320Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/autotune_process.py", line 677, in precompile 2025-12-04T09:54:35.0325769Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] getattr(mod, self.kernel_name).precompile() 2025-12-04T09:54:35.0326269Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/runtime/triton_heuristics.py", line 444, in precompile 2025-12-04T09:54:35.0326725Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] self._make_launchers() 2025-12-04T09:54:35.0327243Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/runtime/triton_heuristics.py", line 613, in _make_launchers 2025-12-04T09:54:35.0327745Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] raise RuntimeError(f"No valid triton configs. {type(exc).__name__}: {exc}") 2025-12-04T09:54:35.0328291Z E1204 09:51:13.364000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] RuntimeError: No valid triton configs. OutOfMemoryError: out of resource: triton_mm Required: 65536 Hardware limit:65536 Reducing block sizes or `num_stages` may help. 2025-12-04T09:54:35.0329390Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] Exception No valid triton configs. OutOfMemoryError: out of resource: triton_mm Required: 65536 Hardware limit:65536 Reducing block sizes or `num_stages` may help. for benchmark choice TritonTemplateCaller(/tmp/tmp8d0mvaed/ug/cugucwjbq4tyf5ajdiyxhb3aaalkdmyvi2nicj4siecpp5ppmd4o.py, ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=128, BLOCK_M=128, BLOCK_N=128, EVEN_K=True, GROUP_M=16, USE_FAST_ACCUM=False, kpack=2, matrix_instr_nonkdim=16, waves_per_eu=0, num_stages=2, num_warps=8) 2025-12-04T09:54:35.0330309Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] Traceback (most recent call last): 2025-12-04T09:54:35.0330756Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/concurrent/futures/thread.py", line 58, in run 2025-12-04T09:54:35.0331161Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] result = self.fn(*self.args, **self.kwargs) 2025-12-04T09:54:35.0331632Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 3255, in precompile_with_captured_stdout 2025-12-04T09:54:35.0332072Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] choice.precompile() 2025-12-04T09:54:35.0332482Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2289, in precompile 2025-12-04T09:54:35.0332901Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] self.bmreq.precompile() 2025-12-04T09:54:35.0333320Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/autotune_process.py", line 677, in precompile 2025-12-04T09:54:35.0333769Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] getattr(mod, self.kernel_name).precompile() 2025-12-04T09:54:35.0334227Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/runtime/triton_heuristics.py", line 444, in precompile 2025-12-04T09:54:35.0334690Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] self._make_launchers() 2025-12-04T09:54:35.0335156Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/runtime/triton_heuristics.py", line 613, in _make_launchers 2025-12-04T09:54:35.0335692Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] raise RuntimeError(f"No valid triton configs. {type(exc).__name__}: {exc}") 2025-12-04T09:54:35.0336254Z E1204 09:51:13.401000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] RuntimeError: No valid triton configs. OutOfMemoryError: out of resource: triton_mm Required: 65536 Hardware limit:65536 Reducing block sizes or `num_stages` may help. 2025-12-04T09:54:35.0337410Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] Exception No valid triton configs. OutOfMemoryError: out of resource: triton_mm Required: 65536 Hardware limit:65536 Reducing block sizes or `num_stages` may help. for benchmark choice TritonTemplateCaller(/tmp/tmp8d0mvaed/os/cosrvtdlho3xj776taejrpguwg66tlbebnarlp7x5o4pyo4xpsjk.py, ACC_TYPE='tl.float32', ALLOW_TF32=False, BLOCK_K=64, BLOCK_M=256, BLOCK_N=256, EVEN_K=True, GROUP_M=4, USE_FAST_ACCUM=False, kpack=2, matrix_instr_nonkdim=16, waves_per_eu=0, num_stages=2, num_warps=8) 2025-12-04T09:54:35.0338317Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] Traceback (most recent call last): 2025-12-04T09:54:35.0338709Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/concurrent/futures/thread.py", line 58, in run 2025-12-04T09:54:35.0339108Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] result = self.fn(*self.args, **self.kwargs) 2025-12-04T09:54:35.0339577Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 3255, in precompile_with_captured_stdout 2025-12-04T09:54:35.0340071Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] choice.precompile() 2025-12-04T09:54:35.0340492Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/select_algorithm.py", line 2289, in precompile 2025-12-04T09:54:35.0340911Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] self.bmreq.precompile() 2025-12-04T09:54:35.0341335Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/autotune_process.py", line 677, in precompile 2025-12-04T09:54:35.0341780Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] getattr(mod, self.kernel_name).precompile() 2025-12-04T09:54:35.0342243Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/runtime/triton_heuristics.py", line 444, in precompile 2025-12-04T09:54:35.0342678Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] self._make_launchers() 2025-12-04T09:54:35.0343116Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/runtime/triton_heuristics.py", line 613, in _make_launchers 2025-12-04T09:54:35.0343615Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] raise RuntimeError(f"No valid triton configs. {type(exc).__name__}: {exc}") 2025-12-04T09:54:35.0344153Z E1204 09:51:13.408000 133810 site-packages/torch/_inductor/select_algorithm.py:3323] RuntimeError: No valid triton configs. OutOfMemoryError: out of resource: triton_mm Required: 65536 Hardware limit:65536 Reducing block sizes or `num_stages` may help. 2025-12-04T09:54:35.0344544Z PASSED [8.8719s] [ 0%] 2025-12-04T09:54:35.0344801Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.4716s] [ 1%] 2025-12-04T09:54:35.0345229Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [1.0047s] [ 2%] 2025-12-04T09:54:35.0345574Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.9974s] [ 2%] 2025-12-04T09:54:35.0345823Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8986s] [ 2%] 2025-12-04T09:54:35.0346066Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.7471s] [ 2%] 2025-12-04T09:54:35.0346307Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.1655s] [ 2%] 2025-12-04T09:54:35.0346551Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8043s] [ 2%] 2025-12-04T09:54:35.0346840Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.7412s] [ 2%] 2025-12-04T09:54:35.0347083Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.7635s] [ 2%] 2025-12-04T09:54:35.0347324Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.7694s] [ 2%] 2025-12-04T09:54:35.0347568Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.9848s] [ 2%] 2025-12-04T09:54:35.0347812Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8357s] [ 2%] 2025-12-04T09:54:35.0348054Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.6198s] [ 2%] 2025-12-04T09:54:35.0348295Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8607s] [ 2%] 2025-12-04T09:54:35.0348589Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8726s] [ 2%] 2025-12-04T09:54:35.0348830Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.9016s] [ 2%] 2025-12-04T09:54:35.0349070Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8674s] [ 2%] 2025-12-04T09:54:35.0349363Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8570s] [ 2%] 2025-12-04T09:54:35.0349609Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8677s] [ 2%] 2025-12-04T09:54:35.0349853Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8678s] [ 2%] 2025-12-04T09:54:35.0350167Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.1008s] [ 2%] 2025-12-04T09:54:35.0350417Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8717s] [ 2%] 2025-12-04T09:54:35.0350663Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8829s] [ 2%] 2025-12-04T09:54:35.0350906Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8769s] [ 2%] 2025-12-04T09:54:35.0351148Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.5980s] [ 2%] 2025-12-04T09:54:35.0351393Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8604s] [ 2%] 2025-12-04T09:54:35.0351636Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.6031s] [ 2%] 2025-12-04T09:54:35.0351883Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8681s] [ 2%] 2025-12-04T09:54:35.0352128Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.6207s] [ 2%] 2025-12-04T09:54:35.0352371Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8785s] [ 2%] 2025-12-04T09:54:35.0352614Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8710s] [ 2%] 2025-12-04T09:54:35.0352857Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.1184s] [ 2%] 2025-12-04T09:54:35.0353121Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8611s] [ 2%] 2025-12-04T09:54:35.0353482Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8682s] [ 2%] 2025-12-04T09:54:35.0353730Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8363s] [ 2%] 2025-12-04T09:54:35.0354014Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8413s] [ 2%] 2025-12-04T09:54:35.0354279Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8423s] [ 2%] 2025-12-04T09:54:35.0354519Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8592s] [ 2%] 2025-12-04T09:54:35.0354759Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.4222s] [ 2%] 2025-12-04T09:54:35.0355001Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8454s] [ 2%] 2025-12-04T09:54:35.0355242Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.0998s] [ 2%] 2025-12-04T09:54:35.0355482Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8494s] [ 2%] 2025-12-04T09:54:35.0355723Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8830s] [ 2%] 2025-12-04T09:54:35.0355968Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.3436s] [ 2%] 2025-12-04T09:54:35.0356214Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [2.8402s] [ 2%] 2025-12-04T09:54:35.0356455Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.6643s] [ 2%] 2025-12-04T09:54:35.0356696Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.3465s] [ 2%] 2025-12-04T09:54:35.0357108Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.4071s] [ 2%] 2025-12-04T09:54:35.0357466Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.4655s] [ 2%] 2025-12-04T09:54:35.0357710Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.3883s] [ 2%] 2025-12-04T09:54:35.0357952Z inductor/test_compile_subprocess.py::GPUTests::test_progressive PASSED [3.6275s] [ 2%] 2025-12-04T09:54:35.0358342Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8151s] [ 2%] 2025-12-04T09:54:35.0358740Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8134s] [ 2%] 2025-12-04T09:54:35.0359134Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8607s] [ 2%] 2025-12-04T09:54:35.0359528Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8104s] [ 2%] 2025-12-04T09:54:35.0359921Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8767s] [ 2%] 2025-12-04T09:54:35.0360313Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8423s] [ 2%] 2025-12-04T09:54:35.0360708Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8598s] [ 2%] 2025-12-04T09:54:35.0361102Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.7963s] [ 2%] 2025-12-04T09:54:35.0361496Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.5521s] [ 2%] 2025-12-04T09:54:35.0361890Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.6172s] [ 2%] 2025-12-04T09:54:35.0362284Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.7064s] [ 2%] 2025-12-04T09:54:35.0362676Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2027s] [ 2%] 2025-12-04T09:54:35.0363086Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.1982s] [ 2%] 2025-12-04T09:54:35.0363502Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2091s] [ 2%] 2025-12-04T09:54:35.0363927Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.7491s] [ 2%] 2025-12-04T09:54:35.0364326Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.7017s] [ 2%] 2025-12-04T09:54:35.0364720Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8054s] [ 2%] 2025-12-04T09:54:35.0365114Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.6032s] [ 2%] 2025-12-04T09:54:35.0365511Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.6044s] [ 2%] 2025-12-04T09:54:35.0365908Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.6036s] [ 2%] 2025-12-04T09:54:35.0366301Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8011s] [ 2%] 2025-12-04T09:54:35.0366693Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.6040s] [ 2%] 2025-12-04T09:54:35.0367179Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8472s] [ 2%] 2025-12-04T09:54:35.0367659Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.7606s] [ 2%] 2025-12-04T09:54:35.0368164Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.5596s] [ 2%] 2025-12-04T09:54:35.0368564Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.6548s] [ 2%] 2025-12-04T09:54:35.0368958Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8677s] [ 2%] 2025-12-04T09:54:35.0369350Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8361s] [ 2%] 2025-12-04T09:54:35.0369745Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8547s] [ 2%] 2025-12-04T09:54:35.0370144Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.8609s] [ 2%] 2025-12-04T09:54:35.0370639Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.6521s] [ 2%] 2025-12-04T09:54:35.0371096Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2129s] [ 2%] 2025-12-04T09:54:35.0371492Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2063s] [ 2%] 2025-12-04T09:54:35.0371886Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2141s] [ 2%] 2025-12-04T09:54:35.0372282Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2101s] [ 2%] 2025-12-04T09:54:35.0372684Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.6039s] [ 2%] 2025-12-04T09:54:35.0373122Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2072s] [ 2%] 2025-12-04T09:54:35.0373560Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2062s] [ 2%] 2025-12-04T09:54:35.0373991Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2108s] [ 2%] 2025-12-04T09:54:35.0374435Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.7343s] [ 2%] 2025-12-04T09:54:35.0374828Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2111s] [ 2%] 2025-12-04T09:54:35.0375223Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2039s] [ 2%] 2025-12-04T09:54:35.0375615Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.3434s] [ 2%] 2025-12-04T09:54:35.0376007Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2148s] [ 2%] 2025-12-04T09:54:35.0376398Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2199s] [ 2%] 2025-12-04T09:54:35.0376879Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2092s] [ 2%] 2025-12-04T09:54:35.0377276Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2096s] [ 2%] 2025-12-04T09:54:35.0377722Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2078s] [ 2%] 2025-12-04T09:54:35.0378118Z inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda <- test/inductor/test_torchinductor.py FAILED [0.2045s] [ 2%] 2025-12-04T09:54:35.0378524Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2549s] [ 2%] 2025-12-04T09:54:35.0378931Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.4958s] [ 2%] 2025-12-04T09:54:35.0379339Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2417s] [ 2%] 2025-12-04T09:54:35.0379744Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2463s] [ 2%] 2025-12-04T09:54:35.0380148Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.5162s] [ 2%] 2025-12-04T09:54:35.0380555Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.4963s] [ 2%] 2025-12-04T09:54:35.0380963Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.4858s] [ 2%] 2025-12-04T09:54:35.0381371Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.4913s] [ 2%] 2025-12-04T09:54:35.0381790Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.4704s] [ 2%] 2025-12-04T09:54:35.0382195Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2437s] [ 2%] 2025-12-04T09:54:35.0382603Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.4963s] [ 2%] 2025-12-04T09:54:35.0383045Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.4832s] [ 2%] 2025-12-04T09:54:35.0383458Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.4995s] [ 2%] 2025-12-04T09:54:35.0383863Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.3663s] [ 2%] 2025-12-04T09:54:35.0384268Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.4803s] [ 2%] 2025-12-04T09:54:35.0384696Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.4673s] [ 2%] 2025-12-04T09:54:35.0385108Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2373s] [ 2%] 2025-12-04T09:54:35.0385521Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2370s] [ 2%] 2025-12-04T09:54:35.0385927Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.7816s] [ 2%] 2025-12-04T09:54:35.0386335Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.4859s] [ 2%] 2025-12-04T09:54:35.0386791Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2459s] [ 2%] 2025-12-04T09:54:35.0387250Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2449s] [ 2%] 2025-12-04T09:54:35.0387659Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.5045s] [ 2%] 2025-12-04T09:54:35.0388067Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.4974s] [ 2%] 2025-12-04T09:54:35.0388470Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.4904s] [ 2%] 2025-12-04T09:54:35.0388884Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2382s] [ 2%] 2025-12-04T09:54:35.0389291Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2397s] [ 2%] 2025-12-04T09:54:35.0389697Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.4756s] [ 2%] 2025-12-04T09:54:35.0390101Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2435s] [ 2%] 2025-12-04T09:54:35.0390522Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2340s] [ 2%] 2025-12-04T09:54:35.0390926Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2425s] [ 2%] 2025-12-04T09:54:35.0391330Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2539s] [ 2%] 2025-12-04T09:54:35.0391733Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.4777s] [ 2%] 2025-12-04T09:54:35.0392139Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2332s] [ 2%] 2025-12-04T09:54:35.0392589Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.5025s] [ 2%] 2025-12-04T09:54:35.0392993Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.4562s] [ 2%] 2025-12-04T09:54:35.0393398Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.4702s] [ 2%] 2025-12-04T09:54:35.0393803Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2575s] [ 2%] 2025-12-04T09:54:35.0394212Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [1.8739s] [ 2%] 2025-12-04T09:54:35.0394621Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.4699s] [ 2%] 2025-12-04T09:54:35.0395030Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.7982s] [ 2%] 2025-12-04T09:54:35.0395435Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.4931s] [ 2%] 2025-12-04T09:54:35.0395841Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2449s] [ 2%] 2025-12-04T09:54:35.0396247Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.4918s] [ 2%] 2025-12-04T09:54:35.0396686Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2234s] [ 2%] 2025-12-04T09:54:35.0397136Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2517s] [ 2%] 2025-12-04T09:54:35.0397543Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.4700s] [ 2%] 2025-12-04T09:54:35.0397960Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py FAILED [0.2378s] [ 2%] 2025-12-04T09:54:35.0398386Z inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda <- test/inductor/test_torchinductor.py PASSED [0.4865s] [ 2%] 2025-12-04T09:54:35.0398620Z 2025-12-04T09:54:35.0398688Z =================================== FAILURES =================================== 2025-12-04T09:54:35.0398906Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0399092Z Traceback (most recent call last): 2025-12-04T09:54:35.0399314Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0399526Z return value(self) 2025-12-04T09:54:35.0399779Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0400034Z self.assertExpectedInline( 2025-12-04T09:54:35.0400311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0400686Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0401019Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0401263Z assert_expected_inline( 2025-12-04T09:54:35.0401515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0401810Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0402117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0402409Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0402647Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0402865Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0403054Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0403240Z raise self.failureException(msg) 2025-12-04T09:54:35.0403444Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0403709Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0404159Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0404615Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0404893Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0405135Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0405454Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0405709Z 2025-12-04T09:54:35.0405791Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0406068Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0406269Z 2025-12-04T09:54:35.0406387Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0406592Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0406839Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0406972Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0407169Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0407661Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0408101Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0408436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0408818Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0409174Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0409526Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0409741Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.0409923Z Traceback (most recent call last): 2025-12-04T09:54:35.0410138Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.0410339Z self.common( 2025-12-04T09:54:35.0410529Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.0410712Z return func(*args, **kwds) 2025-12-04T09:54:35.0410916Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.0411119Z check_model( 2025-12-04T09:54:35.0411298Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.0411494Z assert_equal_fn( 2025-12-04T09:54:35.0411700Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.0411935Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.0412195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.0412466Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.0412644Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.0412732Z 2025-12-04T09:54:35.0412780Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.0412968Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.0413195Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.0413323Z 2025-12-04T09:54:35.0413372Z The failure occurred for item [2] 2025-12-04T09:54:35.0413450Z 2025-12-04T09:54:35.0413528Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0413792Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.0413984Z 2025-12-04T09:54:35.0414089Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0414330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0414502Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.0414715Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.0415188Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.0415618Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0415957Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0416292Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0416635Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0417020Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0417358Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0417692Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0418033Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0418367Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0418574Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0418765Z Traceback (most recent call last): 2025-12-04T09:54:35.0418967Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0419194Z return value(self) 2025-12-04T09:54:35.0419411Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0419662Z self.assertExpectedInline( 2025-12-04T09:54:35.0419914Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0420263Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0420583Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0420817Z assert_expected_inline( 2025-12-04T09:54:35.0421038Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0421275Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0421526Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0421807Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0422037Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0422252Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0422440Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0422618Z raise self.failureException(msg) 2025-12-04T09:54:35.0472735Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0472988Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0473471Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0473926Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0474198Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0474436Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0474753Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0474969Z 2025-12-04T09:54:35.0475043Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0475304Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0475491Z 2025-12-04T09:54:35.0475579Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0475779Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0475933Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0476065Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0476261Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0476727Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0477231Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0477556Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0477939Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0478272Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0478605Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0478801Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0478956Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0479086Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0479274Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0479735Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0480168Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0480490Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0480819Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0481149Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0481476Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0481722Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0481887Z Traceback (most recent call last): 2025-12-04T09:54:35.0482085Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0482277Z return value(self) 2025-12-04T09:54:35.0482488Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0482722Z self.assertExpectedInline( 2025-12-04T09:54:35.0482971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0483314Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0483629Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0483859Z assert_expected_inline( 2025-12-04T09:54:35.0484081Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0484312Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0484563Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0484836Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0485061Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0485274Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0485458Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0485633Z raise self.failureException(msg) 2025-12-04T09:54:35.0485806Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0486055Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0486510Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0487131Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0487401Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0487630Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0487940Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0488151Z 2025-12-04T09:54:35.0488227Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0488483Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0488670Z 2025-12-04T09:54:35.0488756Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0488953Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0489104Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0489231Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0489422Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0489921Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0490352Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0490677Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0491010Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0491343Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0491676Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0491875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0492028Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0492157Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0492350Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0492808Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0493240Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0493558Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0493890Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0494220Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0494552Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0494766Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0494936Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0495064Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0495254Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0495714Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0496143Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0496465Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0496849Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0497180Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0497514Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0497720Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0497881Z Traceback (most recent call last): 2025-12-04T09:54:35.0498079Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0498270Z return value(self) 2025-12-04T09:54:35.0498482Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0498712Z self.assertExpectedInline( 2025-12-04T09:54:35.0501436Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0501782Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0502099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0502324Z assert_expected_inline( 2025-12-04T09:54:35.0502539Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0502772Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0503018Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0503289Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0503517Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0503731Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0503913Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0504091Z raise self.failureException(msg) 2025-12-04T09:54:35.0504263Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0504510Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0504937Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0505382Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0512091Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0512343Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0512651Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0512863Z 2025-12-04T09:54:35.0512937Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0513194Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0513378Z 2025-12-04T09:54:35.0513464Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0513663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0513816Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0513946Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0514139Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0514597Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0515026Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0515350Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0515683Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0516045Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0516379Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0516579Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0516730Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0516889Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0517078Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0517534Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0517966Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0518290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0518627Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0518957Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0519289Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0519486Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0519637Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0519774Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0519963Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0520423Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0520900Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0521219Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0521549Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0521880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0522208Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0522406Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0522560Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0522687Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0522878Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0523335Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0523760Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0524081Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0524418Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0524786Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0525129Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0525340Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0525505Z Traceback (most recent call last): 2025-12-04T09:54:35.0525701Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0525892Z return value(self) 2025-12-04T09:54:35.0526102Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0526332Z self.assertExpectedInline( 2025-12-04T09:54:35.0526582Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0526970Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0527289Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0527521Z assert_expected_inline( 2025-12-04T09:54:35.0527743Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0527980Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0528233Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0528510Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0528739Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0528957Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0529146Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0529348Z raise self.failureException(msg) 2025-12-04T09:54:35.0529544Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0529795Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0530231Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0530669Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0530939Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0531170Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0531485Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0531697Z 2025-12-04T09:54:35.0531772Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0532036Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0532221Z 2025-12-04T09:54:35.0532312Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0532514Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0532674Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0532807Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0533036Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0533501Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0533936Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0534268Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0534607Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0534945Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0535284Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0535488Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0535644Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0535774Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0535966Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0536419Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0536892Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0537215Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0537547Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0537898Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0538251Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0538451Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0538602Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0538729Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0538918Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0539379Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0539809Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0540131Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0540467Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0540804Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0541137Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0541332Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0541483Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0541638Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0541826Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0542284Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0542713Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0543032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0543363Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0543696Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0544030Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0544230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0544383Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0544510Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0544701Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0545159Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0545588Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0545910Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0546261Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0546612Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0547022Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0547231Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0547395Z Traceback (most recent call last): 2025-12-04T09:54:35.0547590Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0547786Z return value(self) 2025-12-04T09:54:35.0548002Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0548236Z self.assertExpectedInline( 2025-12-04T09:54:35.0548491Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0548838Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0549161Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0549395Z assert_expected_inline( 2025-12-04T09:54:35.0549615Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0549846Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0550096Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0550369Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0550645Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0550864Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0551050Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0551228Z raise self.failureException(msg) 2025-12-04T09:54:35.0551402Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0551651Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0552081Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0552528Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0552801Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0553033Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0553340Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0553560Z 2025-12-04T09:54:35.0553636Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0553895Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0554084Z 2025-12-04T09:54:35.0554172Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0554370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0554526Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0554687Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0554903Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0555365Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0555795Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0556122Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0556458Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0556861Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0557194Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0557395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0557546Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0557674Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0557864Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0558321Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0558788Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0559112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0559449Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0559780Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0560113Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0560313Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0560464Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0560591Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0560783Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0561241Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0561676Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0561996Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0562325Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0562661Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0562993Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0563193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0563364Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0563497Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0563702Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0564159Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0564590Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0564907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0565238Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0565574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0565911Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0566113Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0566268Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0566398Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0566589Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0567085Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0567548Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0567878Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0568221Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0568561Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0568900Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0569103Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0569259Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0569390Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0569584Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0570046Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0570480Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0570801Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0571133Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0571469Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0571802Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0572013Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0572199Z Traceback (most recent call last): 2025-12-04T09:54:35.0572400Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0572614Z return value(self) 2025-12-04T09:54:35.0572827Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0573058Z self.assertExpectedInline( 2025-12-04T09:54:35.0573307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0573650Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0573972Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0574206Z assert_expected_inline( 2025-12-04T09:54:35.0574425Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0574662Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0574913Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0575187Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0575412Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0575628Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0575813Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0575995Z raise self.failureException(msg) 2025-12-04T09:54:35.0576170Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0576442Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0576951Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0577395Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0577663Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0577894Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0578208Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0578423Z 2025-12-04T09:54:35.0578500Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0578758Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0578943Z 2025-12-04T09:54:35.0579034Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0579232Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0579387Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0579518Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0579714Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0580178Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0580632Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0580985Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0581325Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0581665Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0582005Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0582209Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0582366Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0582501Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0582693Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0583166Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0583600Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0583926Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0584264Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0584638Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0584975Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0585178Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0585335Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0585466Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0585661Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0586120Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0586555Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0586924Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0587265Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0587599Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0587936Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0588136Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0588292Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0588423Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0588615Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0589078Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0589525Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0589860Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0590196Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0590532Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0590866Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0591064Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0591216Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0591346Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0591544Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0592005Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0592435Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0592756Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0593088Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0593502Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0593836Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0594035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0594187Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0594314Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0594502Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0594955Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0595382Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0595701Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0596032Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0596364Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0596694Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0596950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0597100Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0597225Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0597415Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0597874Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0598325Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0598666Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0599004Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0599339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0599667Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0599870Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0600033Z Traceback (most recent call last): 2025-12-04T09:54:35.0600230Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0600425Z return value(self) 2025-12-04T09:54:35.0600644Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0600880Z self.assertExpectedInline( 2025-12-04T09:54:35.0601137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0601483Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0601802Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0602029Z assert_expected_inline( 2025-12-04T09:54:35.0602280Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0602514Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0602765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0603038Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0603266Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0603479Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0603663Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0603839Z raise self.failureException(msg) 2025-12-04T09:54:35.0604011Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0604263Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0607028Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0607481Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0607756Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0607990Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0608305Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0608517Z 2025-12-04T09:54:35.0608594Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0608851Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0609078Z 2025-12-04T09:54:35.0609164Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0609361Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0609513Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0609639Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0609830Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0610289Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0610719Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0611047Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0611384Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0611717Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0612053Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0612257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0612416Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0612551Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0612774Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0613238Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0613678Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0614003Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0614340Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0614679Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0615016Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0615222Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0615380Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0615508Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0615698Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0616158Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0616589Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0617024Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0617365Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0617720Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0618073Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0618276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0618430Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0618562Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0618754Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0619212Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0619643Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0619970Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0620307Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0620641Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0620976Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0621179Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0621334Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0621466Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0631078Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0631545Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0631978Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0632299Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0632630Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0632962Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0633297Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0633495Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0633650Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0633780Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0633972Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0634433Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0634861Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0635183Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0635519Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0635878Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0636228Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0636474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0636703Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0636909Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0637132Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0637643Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0638115Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0638469Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0638865Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0639238Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0639632Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0639869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0640052Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0640233Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0647022Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0647562Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0648041Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0648423Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0648801Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0649171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0649559Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0649822Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0650041Z Traceback (most recent call last): 2025-12-04T09:54:35.0650271Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0650505Z return value(self) 2025-12-04T09:54:35.0650763Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0651037Z self.assertExpectedInline( 2025-12-04T09:54:35.0651328Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0651739Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0652099Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0652374Z assert_expected_inline( 2025-12-04T09:54:35.0652663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0652943Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0653249Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0653574Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0653831Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0654100Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0654319Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0654544Z raise self.failureException(msg) 2025-12-04T09:54:35.0654754Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0655041Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0655540Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0656015Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0656313Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0656606Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0657069Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0657294Z 2025-12-04T09:54:35.0657414Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0657714Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0657910Z 2025-12-04T09:54:35.0658029Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0658297Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0658486Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0658648Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0658895Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0659397Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0659875Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0660243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0660614Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0661126Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0661496Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0661723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0661928Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0662101Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0662354Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0662873Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0663333Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0663694Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0664066Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0664434Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0664830Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0665067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0665256Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0665429Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0665651Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0666163Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0666632Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0667089Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0667478Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0667847Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0668227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0668462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0668655Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0668829Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0669058Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0669554Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0670056Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0670419Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0670802Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0671179Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0671546Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0671798Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0671992Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0672206Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0672440Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0672945Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0673427Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0690548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0690910Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0691316Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0691688Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0691938Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0692133Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0692296Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0692541Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0693034Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0693495Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0693916Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0694289Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0694679Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0695051Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0695290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0695495Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0695660Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0695884Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0696402Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0696901Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0697261Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0697638Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0698023Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0698404Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0698643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0698822Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0699029Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0699276Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0699775Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0700246Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0700601Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0700985Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0701357Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0701720Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0701980Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0702166Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0702335Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0702567Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0703054Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0749299Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0749750Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0750107Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0750451Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0750788Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0751000Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0751170Z Traceback (most recent call last): 2025-12-04T09:54:35.0751378Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0751575Z return value(self) 2025-12-04T09:54:35.0751793Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0752033Z self.assertExpectedInline( 2025-12-04T09:54:35.0752292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0752647Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0752974Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0753207Z assert_expected_inline( 2025-12-04T09:54:35.0753429Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0753666Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0753923Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0754202Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0754461Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0754708Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0754895Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0755073Z raise self.failureException(msg) 2025-12-04T09:54:35.0755248Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0755503Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0755940Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0756393Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0756666Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0756943Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0757257Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0757470Z 2025-12-04T09:54:35.0757551Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0757813Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0758002Z 2025-12-04T09:54:35.0758110Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0758333Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0758493Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0758626Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0758827Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0759295Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0759734Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0760066Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0760408Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0760746Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0761083Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0761281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0761435Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0761565Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0761758Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0762219Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0762667Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0763008Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0763349Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0763686Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0764018Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0764218Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0764371Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0764499Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0764693Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0765159Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0765590Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0765909Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0766246Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0766580Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0766983Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0767184Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0767336Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0767464Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0767654Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0768109Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0768538Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0768863Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0769200Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0769534Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0769868Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0770065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0770217Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0770344Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0770534Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0770997Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0771442Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0771789Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0772122Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0772453Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0772785Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0772983Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0773137Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0773198Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0773299Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0773624Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0773702Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0773912Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0773999Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0774224Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0774335Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0774409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0774456Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0774512Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0774613Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0774932Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0775008Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0775219Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0775310Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0775518Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0775607Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0775680Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0775724Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0775780Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0775882Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0776204Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0776294Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0776528Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0776617Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0776872Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0776958Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0777031Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0777073Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0777133Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0777235Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0777555Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0777627Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0777836Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0777923Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0778143Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0778247Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0778323Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0778367Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0778424Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0778525Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0778846Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0778919Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0779133Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0779222Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0779429Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0779517Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0779600Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0779650Z Traceback (most recent call last): 2025-12-04T09:54:35.0779773Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0779815Z return value(self) 2025-12-04T09:54:35.0779967Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0780017Z self.assertExpectedInline( 2025-12-04T09:54:35.0780195Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0780367Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0780515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0780560Z assert_expected_inline( 2025-12-04T09:54:35.0780707Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0780760Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0780928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0781006Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0781126Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0781191Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0781284Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0781335Z raise self.failureException(msg) 2025-12-04T09:54:35.0781426Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0781549Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0781826Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0781986Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0782091Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0782191Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0782373Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0782375Z 2025-12-04T09:54:35.0782451Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0782602Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0782604Z 2025-12-04T09:54:35.0782692Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0782770Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0782815Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0782874Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0782977Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0783298Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0783372Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0783585Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0783676Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0783886Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0784002Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0784075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0784118Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0784175Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0784275Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0784594Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0784670Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0784881Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0784972Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0785179Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0785267Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0785338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0785381Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0785438Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0785547Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0785877Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0785954Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0786163Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0786255Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0786466Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0786552Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0786627Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0786670Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0786728Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0786867Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0787185Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0787257Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0787465Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0787552Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0787760Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0787881Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0787955Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0787997Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0788055Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0788153Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0788474Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0788549Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0788758Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0788847Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0789052Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0789140Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0789212Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0789256Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0789312Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0789433Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0789766Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0789843Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0790053Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0790143Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0790352Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0790441Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0790516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0790561Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0790619Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0790722Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0791045Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0791120Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0791329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0791418Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0791639Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0791737Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0791811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0791852Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0791909Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0792008Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0792326Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0792400Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0792610Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0792698Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0792908Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0792995Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0793068Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0793110Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0793168Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0793279Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0793613Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0793689Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0793898Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0793987Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0794195Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0794285Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0794359Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0794409Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0794468Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0794570Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0794887Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0794963Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0795174Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0795267Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0795492Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0795596Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0795668Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0795713Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0795770Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0795871Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0796194Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0796269Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0796482Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0796570Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0796830Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0796921Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0797005Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0797053Z Traceback (most recent call last): 2025-12-04T09:54:35.0797176Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0797237Z return value(self) 2025-12-04T09:54:35.0797411Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0797463Z self.assertExpectedInline( 2025-12-04T09:54:35.0797641Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0797781Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0797932Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0797977Z assert_expected_inline( 2025-12-04T09:54:35.0798127Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0798179Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0798349Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0798425Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0798546Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0798609Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0798705Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0798753Z raise self.failureException(msg) 2025-12-04T09:54:35.0798843Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0798963Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0799239Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0799423Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0799522Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0799620Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0799799Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0799801Z 2025-12-04T09:54:35.0799879Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0800030Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0800034Z 2025-12-04T09:54:35.0800124Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0800199Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0800244Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0800300Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0800402Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0800723Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0800798Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0801039Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0801129Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0801343Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0801430Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0801506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0801550Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0801606Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0801707Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0802030Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0802106Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0802320Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0802408Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0802619Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0802707Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0802781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0802825Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0802895Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0802993Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0803338Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0803412Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0803621Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0803709Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0803918Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0804006Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0804079Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0804122Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0804179Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0804282Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0804600Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0804677Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0804909Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0805001Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0805208Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0805297Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0805371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0805414Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0805471Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0805573Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0805896Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0805971Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0806184Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0806270Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0806482Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0806570Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0806646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0806689Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0806802Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0806903Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0807241Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0807315Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0807527Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0807617Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0807827Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0807915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0807991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0808032Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0808092Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0808192Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0808515Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0808589Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0808824Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0808915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0809124Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0809214Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0809285Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0809331Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0809386Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0809487Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0809804Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0809881Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0810092Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0810183Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0810390Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0810479Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0810554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0810598Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0810672Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0810783Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0811104Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0811178Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0811391Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0811478Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0811691Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0811780Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0811856Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0811899Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0811959Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0812058Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0812378Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0812461Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0812712Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0812803Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0813012Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0813102Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0813174Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0813218Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0813276Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0813376Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0813697Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0813773Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0813981Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0814069Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0814277Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0814365Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0814438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0814495Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0814551Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0814663Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0814980Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0815054Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0815261Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0815349Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0815557Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0815644Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0815726Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0815774Z Traceback (most recent call last): 2025-12-04T09:54:35.0815900Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0815938Z return value(self) 2025-12-04T09:54:35.0816090Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0816140Z self.assertExpectedInline( 2025-12-04T09:54:35.0816318Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0816474Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0816624Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0816668Z assert_expected_inline( 2025-12-04T09:54:35.0817073Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0817125Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0817290Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0817366Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0817487Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0817548Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0817645Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0817693Z raise self.failureException(msg) 2025-12-04T09:54:35.0817787Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0817907Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0818183Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0818317Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0818419Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0818545Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0818735Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0818738Z 2025-12-04T09:54:35.0818814Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0818962Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0818964Z 2025-12-04T09:54:35.0819053Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0819127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0819170Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0819228Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0819333Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0819657Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0819733Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0819946Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0820036Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0820244Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0820359Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0820435Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0820478Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0820535Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0820634Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0820954Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0821026Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0821236Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0821325Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0821537Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0821624Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0821702Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0821744Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0821805Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0821903Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0822226Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0822310Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0822530Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0822616Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0822824Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0822911Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0822983Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0823025Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0823082Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0823183Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0823507Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0823580Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0823790Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0823880Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0824100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0824203Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0824278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0824322Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0824378Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0824477Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0824793Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0824866Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0825074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0825164Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0825373Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0825460Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0825535Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0825576Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0825634Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0825732Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0826054Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0830578Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0833915Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0834001Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0834209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0834296Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0834369Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0834411Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0834471Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0834573Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0834896Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0834969Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0835183Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0835272Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0843756Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0843870Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0843946Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0843990Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0844047Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0844147Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0844463Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0844540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0844753Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0844843Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0845050Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0845136Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0845207Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0845252Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0845309Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0845411Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0845730Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0845822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0846048Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0846135Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0846341Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0846427Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0846500Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0846541Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0846601Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0846699Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0847052Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0847125Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0847335Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0847420Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0847667Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0847756Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0847831Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0847872Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0847929Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0848028Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0848348Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0848421Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0848630Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0848718Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0848926Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0849015Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0849089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0849134Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0849192Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0849294Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0849614Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0849719Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0849927Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0850014Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0850222Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0850310Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0850382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0850427Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0850489Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0850592Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0850918Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0850991Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0851204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0851292Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0851525Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0851613Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0851698Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0851744Z Traceback (most recent call last): 2025-12-04T09:54:35.0851868Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0851906Z return value(self) 2025-12-04T09:54:35.0852060Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0852109Z self.assertExpectedInline( 2025-12-04T09:54:35.0852286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0852425Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0852576Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0852621Z assert_expected_inline( 2025-12-04T09:54:35.0852770Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0852822Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0852986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0853062Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0853181Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0853242Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0853338Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0853386Z raise self.failureException(msg) 2025-12-04T09:54:35.0853490Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0853619Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0853894Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0854030Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0854130Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0854231Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0854410Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0854413Z 2025-12-04T09:54:35.0854489Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0854637Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0854639Z 2025-12-04T09:54:35.0854728Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0854802Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0854847Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0854903Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0855016Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0855353Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0855430Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0855643Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0855731Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0855943Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0856031Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0856109Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0856153Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0856213Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0856312Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0856631Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0856706Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0856953Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0857043Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0857274Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0857378Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0857452Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0857493Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0857550Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0857649Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0857970Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0858046Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0858254Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0858343Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0858547Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0858634Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0858706Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0858751Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0858809Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0858926Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0859259Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0859337Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0859544Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0859633Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0859840Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0859931Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0860007Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0860053Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0860111Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0860212Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0860534Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0860607Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0860817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0860907Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0861128Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0861229Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0861303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0861344Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0861405Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0861504Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0861825Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0861900Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0862112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0862200Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0862410Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0862497Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0862571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0862613Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0862670Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0862799Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0863118Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0863193Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0863403Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0863494Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0863703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0863793Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0863867Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0863914Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0863969Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0864068Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0864390Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0864468Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0864679Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0864770Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0864993Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0865093Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0865169Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0865210Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0865268Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0865367Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0865690Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0865765Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0865976Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0866062Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0866271Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0866356Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0866431Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0866474Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0866534Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0866656Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0867010Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0867084Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0867295Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0867384Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0867591Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0867680Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0867752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0867796Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0867851Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0867952Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0868267Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0868341Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0868550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0868661Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0868868Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0868975Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0869048Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0869092Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0869147Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0869248Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0869566Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0869643Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0869854Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0869942Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0870150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0870237Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0870311Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0870353Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0870422Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0870537Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0870859Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0870932Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0871144Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0871232Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0871442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0871529Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0871602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0871645Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0871702Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0871801Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0872121Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0872194Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0872404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0872507Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0872723Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0872812Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0872892Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0872939Z Traceback (most recent call last): 2025-12-04T09:54:35.0873060Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0873099Z return value(self) 2025-12-04T09:54:35.0873250Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0873299Z self.assertExpectedInline( 2025-12-04T09:54:35.0873478Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0873619Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0873768Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0873814Z assert_expected_inline( 2025-12-04T09:54:35.0873961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0874016Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0874181Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0874261Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0874399Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0874465Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0874562Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0874614Z raise self.failureException(msg) 2025-12-04T09:54:35.0874707Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0874831Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0875104Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0875243Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0875347Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0875446Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0875627Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0875630Z 2025-12-04T09:54:35.0875706Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0875858Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0875861Z 2025-12-04T09:54:35.0875949Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0876028Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0876086Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0876146Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0876262Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0876585Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0876661Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0876921Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0877012Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0877226Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0877320Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0877394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0877439Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0877495Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0877597Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0877916Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0878007Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0878231Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0878325Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0878534Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0878626Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0878699Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0878748Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0878805Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0878907Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0879232Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0879309Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0879518Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0879608Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0879818Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0879906Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0879983Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0880041Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0880102Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0880213Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0880532Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0880607Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0880818Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0880906Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0881120Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0881209Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0881287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0881330Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0881391Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0881492Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0881813Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0881899Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0882121Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0882211Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0882422Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0882514Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0882588Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0882633Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0882691Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0882793Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0883114Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0883192Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0883401Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0883490Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0883698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0883790Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0883866Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0883922Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0883981Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0884093Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0884410Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0884488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0884700Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0884788Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0884999Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0885087Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0885163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0885205Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0885261Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0885360Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0885679Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0885775Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0885988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0886077Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0886290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0886378Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0886455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0886497Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0886554Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0886656Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0887022Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0887103Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0887312Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0887403Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0887610Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0887701Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0887776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0887836Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0887907Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0888009Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0888327Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0888404Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0888614Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0888708Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0888919Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0889010Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0889085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0889129Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0889188Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0889289Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0889610Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0889711Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0889926Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0890016Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0890226Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0890313Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0890390Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0890434Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0890495Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0890595Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0890919Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0890994Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0891208Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0891295Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0891505Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0891592Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0891669Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0891735Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0891801Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0891905Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0892226Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0892304Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0892513Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0892604Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0892812Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0892904Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0892978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0893023Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0893082Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0893186Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0893522Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0893611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0893823Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0893914Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0894123Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0894214Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0894290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0894333Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0894393Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0894494Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0894815Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0894890Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0895103Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0895193Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0895406Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0895492Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0895596Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0895655Z Traceback (most recent call last): 2025-12-04T09:54:35.0895786Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0895828Z return value(self) 2025-12-04T09:54:35.0895984Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0896036Z self.assertExpectedInline( 2025-12-04T09:54:35.0896220Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0896359Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0896512Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0896558Z assert_expected_inline( 2025-12-04T09:54:35.0896710Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0896797Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0896968Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0897047Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0897170Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0897238Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0897334Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0897387Z raise self.failureException(msg) 2025-12-04T09:54:35.0897493Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0897633Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0897912Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0898056Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0898158Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0898263Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0898444Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0898447Z 2025-12-04T09:54:35.0898531Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0898684Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0898686Z 2025-12-04T09:54:35.0898779Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0898856Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0898906Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0898966Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0899072Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0899403Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0899504Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0899723Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0899814Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0900029Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0900115Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0900194Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0900240Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0900304Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0900406Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0900731Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0900808Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0901024Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0901113Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0901348Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0901444Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0901520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0901569Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0901629Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0901738Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0902060Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0902140Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0902355Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0902451Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0902666Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0902760Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0902836Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0902885Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0902944Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0903050Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0903371Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0903471Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0903682Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0903775Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0903989Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0904079Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0904159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0904204Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0904268Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0904367Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0904690Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0904762Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0904974Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0905061Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0905297Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0905387Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0905465Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0905507Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0905572Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0905671Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0905997Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0906072Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0906287Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0906380Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0906590Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0906680Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0907090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0907135Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0907191Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0907293Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0907614Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0907722Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0907932Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0908022Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0908230Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0908320Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0908393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0908440Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0908500Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0908605Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0908923Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0908999Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0909211Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0909299Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0909543Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0909633Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0909708Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0909749Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0909809Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0909908Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0910228Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0910301Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0910512Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0910601Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0910809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0910896Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0910970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0911014Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0911072Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0911172Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0911494Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0911593Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0911805Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0911898Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0912105Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0912196Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0912271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0912317Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0912374Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0912474Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0912792Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0912868Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0913082Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0913173Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0924463Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0924560Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0924633Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0924681Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0924738Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0924839Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0925166Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0925242Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0925455Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0925545Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0925752Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0925838Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0925912Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0925955Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0926013Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0926111Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0926431Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0926540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0926794Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0926881Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0927090Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0927177Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0927257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0927303Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0927361Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0927461Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0927781Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0927855Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0928062Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0928168Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0928387Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0928476Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0928548Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0928593Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0928648Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0928749Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0929067Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0929148Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0929358Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0929448Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0929656Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0929745Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0929817Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0929861Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0929917Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0930017Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0930338Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0930444Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0930654Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0930741Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0930953Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0931039Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0931122Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0931170Z Traceback (most recent call last): 2025-12-04T09:54:35.0931294Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0931333Z return value(self) 2025-12-04T09:54:35.0931486Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0931535Z self.assertExpectedInline( 2025-12-04T09:54:35.0931712Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0931852Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0932002Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0932064Z assert_expected_inline( 2025-12-04T09:54:35.0932226Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0932279Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0932445Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0932520Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0932638Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0932699Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0932793Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0932841Z raise self.failureException(msg) 2025-12-04T09:54:35.0932932Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0933054Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0933332Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0933470Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0933568Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0933666Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0933845Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0933861Z 2025-12-04T09:54:35.0933937Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0934096Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0934099Z 2025-12-04T09:54:35.0934188Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0934263Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0934307Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0934363Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0934464Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0934787Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0934864Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0935079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0935168Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0935380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0935466Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0935540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0935582Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0935653Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0935763Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0936083Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0936158Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0936368Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0936454Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0936664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0936790Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0936865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0936909Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0936967Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0937067Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0937392Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0937467Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0937678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0937787Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0938010Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0938099Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0938171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0938214Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0938271Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0938371Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0938690Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0938765Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0938973Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0939062Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0939270Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0939358Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0939430Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0939473Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0939556Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0939657Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0939983Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0940055Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0940266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0940352Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0940560Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0940647Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0940722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0940764Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0940821Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0940919Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0941236Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0941309Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0941519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0941617Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0941836Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0941924Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0941996Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0942040Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0942096Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0942195Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0942516Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0942592Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0942800Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0942887Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0943094Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0943181Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0943253Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0943307Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0943373Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0943473Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0943791Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0943866Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0944075Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0944163Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0944372Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0944459Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0944533Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0944575Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0944633Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0944730Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0945048Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0945120Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0945330Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0945428Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0945656Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0945742Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0945815Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0945856Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0945913Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0946010Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0946340Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0946414Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0946624Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0946712Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0946956Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0947043Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0947115Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0947175Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0947247Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0947347Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0947665Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0947739Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0947948Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0948035Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0948243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0948330Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0948403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0948446Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0948502Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0948603Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0948924Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0972191Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0972469Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0972649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0972865Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0972957Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0973038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0973085Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0973148Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0973255Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0973590Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0973676Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0973893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0973985Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0974197Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0974285Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0974379Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0974438Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0974502Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0974607Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0974929Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0975005Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0975219Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0975311Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0975522Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0975616Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0975690Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0975736Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0975794Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0975897Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0976215Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0976294Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0976504Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0976618Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0976862Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0976953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0977026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0977073Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0977131Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0977232Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0977559Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0977638Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0977851Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0977939Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0978154Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0978241Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0978330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0978388Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0978447Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0978549Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0978869Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0978943Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0979157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0979250Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0979473Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.0979564Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.0979652Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.0979706Z Traceback (most recent call last): 2025-12-04T09:54:35.0979837Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.0979878Z return value(self) 2025-12-04T09:54:35.0980035Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.0980088Z self.assertExpectedInline( 2025-12-04T09:54:35.0980271Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.0980413Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.0980581Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.0980644Z assert_expected_inline( 2025-12-04T09:54:35.0980793Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.0980848Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.0981014Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.0981092Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.0981211Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.0981274Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.0981371Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.0981424Z raise self.failureException(msg) 2025-12-04T09:54:35.0981520Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.0981645Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.0981924Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.0982066Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.0982177Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.0982290Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.0982475Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.0982480Z 2025-12-04T09:54:35.0982558Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.0982711Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.0982714Z 2025-12-04T09:54:35.0982805Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.0982884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.0982928Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.0982988Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.0983092Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.0983418Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.0983494Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.0983712Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.0983803Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1004596Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1004700Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1004795Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1004852Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1004911Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1005011Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1005334Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1005409Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1005622Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1005713Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1005921Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1006009Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1006081Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1006124Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1006180Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1006280Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1006648Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1006723Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1006969Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1007058Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1007264Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1007350Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1007421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1007464Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1007520Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1007621Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1007941Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1008016Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1008224Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1008311Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1008520Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1008607Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1008697Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1008753Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1008809Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1008907Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1009222Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1009294Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1009502Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1009591Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1009803Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1009890Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1009962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1010002Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1010058Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1010155Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1010496Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1010571Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1010782Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1010870Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1011076Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1011163Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1011235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1011277Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1011333Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1011434Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1011749Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1011824Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1012032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1012119Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1012324Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1012428Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1012499Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1012555Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1012610Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1012708Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1013026Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1013097Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1013308Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1013395Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1013603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1013689Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1013761Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1013802Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1013857Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1013954Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1014297Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1014371Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1014580Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1014666Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1014871Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1014956Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1015029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1015069Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1015126Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1015224Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1015539Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1015611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1015819Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1015906Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1016111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1016212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1016300Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1016344Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1016399Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1016499Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1016857Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1016929Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1017140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1017227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1017436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1017522Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1017595Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1017638Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1017695Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1017792Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1018145Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1018219Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1018428Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1018514Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1018722Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1018808Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1018881Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1018924Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1018981Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1019080Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1019397Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1019472Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1019679Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1019766Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1019977Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1020084Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1020178Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1020222Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1020277Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1020376Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1020692Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1020765Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1020974Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1021064Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1021272Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1021360Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1021433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1021475Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1021532Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1021630Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1021970Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1022046Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1022257Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1022344Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1022551Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1022637Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1022711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1022753Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1022812Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1022909Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1023227Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1023299Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1023509Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1023597Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1023805Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1023907Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1023990Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1024035Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1024091Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1024192Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1024514Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1024588Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1024799Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1024890Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1025098Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1025187Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1025259Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1025302Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1025357Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1025457Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1025798Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1025876Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1026084Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1026174Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1026383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1026469Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1026553Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1026603Z Traceback (most recent call last): 2025-12-04T09:54:35.1026732Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1026803Z return value(self) 2025-12-04T09:54:35.1026958Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1027009Z self.assertExpectedInline( 2025-12-04T09:54:35.1027188Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1027325Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1027475Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1027520Z assert_expected_inline( 2025-12-04T09:54:35.1027672Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1027744Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1027928Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1028005Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1028125Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1028188Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1028284Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1028334Z raise self.failureException(msg) 2025-12-04T09:54:35.1028426Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1028548Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1028835Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1028976Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1029079Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1029178Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1029358Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1029376Z 2025-12-04T09:54:35.1029470Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1029621Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1029625Z 2025-12-04T09:54:35.1029715Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1029790Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1029834Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1029891Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1029992Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1030314Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1030391Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1030604Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1030696Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1030906Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1030995Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1031070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1031112Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1031169Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1031270Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1034614Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1038605Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1038814Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1038902Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1039110Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1039197Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1039274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1039316Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1039373Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1039470Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1039788Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1039862Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1040073Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1045484Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1045695Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1045785Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1045857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1045900Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1045956Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1046055Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1046374Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1046454Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1046663Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1046780Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1046987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1047075Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1047147Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1047190Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1047246Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1047349Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1047695Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1047790Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1047999Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1048088Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1048297Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1048386Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1048461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1048504Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1048561Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1048660Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1048978Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1049051Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1049276Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1049377Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1049588Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1049675Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1049750Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1049792Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1049848Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1049946Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1050265Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1050337Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1050548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1050637Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1050843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1050930Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1051002Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1051048Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1051105Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1051206Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1051535Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1051624Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1051832Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1051921Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1052129Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1052220Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1052293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1052338Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1052393Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1052494Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1052811Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1052887Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1053111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1053211Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1053420Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1053506Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1053581Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1053622Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1053680Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1053777Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1054097Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1054170Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1054382Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1054469Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1054677Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1054764Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1054837Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1054878Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1054936Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1055048Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1055375Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1055449Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1055657Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1055745Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1055954Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1056045Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1056117Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1056163Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1056218Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1056318Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1056634Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1056708Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1056983Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1057073Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1057280Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1057367Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1057439Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1057485Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1057540Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1057639Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1057962Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1058039Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1058251Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1058337Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1058543Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1058633Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1058706Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1058747Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1058804Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1058929Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1059261Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1059337Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1059546Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1059635Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1059847Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1059937Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1060014Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1060056Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1060116Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1060216Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1060543Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1060617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1060871Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1060962Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1061175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1061262Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1061337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1061379Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1061437Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1061537Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1061861Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1061937Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1062150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1062244Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1062457Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1062549Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1062623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1062669Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1062726Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1062842Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1063171Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1063248Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1063457Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1063549Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1063757Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1063849Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1063922Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1063969Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1064026Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1064126Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1064445Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1064523Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1064761Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1064850Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1065063Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1065152Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1065227Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1065271Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1065330Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1065428Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1065748Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1065823Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1066034Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1066121Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1066333Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1066420Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1066506Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1066557Z Traceback (most recent call last): 2025-12-04T09:54:35.1066698Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1066786Z return value(self) 2025-12-04T09:54:35.1066944Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1066994Z self.assertExpectedInline( 2025-12-04T09:54:35.1067175Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1067311Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1067463Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1067511Z assert_expected_inline( 2025-12-04T09:54:35.1067661Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1067717Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1067881Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1067964Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1068082Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1068148Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1068242Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1068292Z raise self.failureException(msg) 2025-12-04T09:54:35.1068384Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1068522Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1068812Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1068953Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1069053Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1069157Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1069335Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1069338Z 2025-12-04T09:54:35.1069417Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1069570Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1069573Z 2025-12-04T09:54:35.1069663Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1069741Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1069784Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1069844Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1069945Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1070266Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1070342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1070574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1070678Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1070892Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1070981Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1071059Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1071103Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1071163Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1071262Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1071589Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1071664Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1071880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1071971Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1072181Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1072284Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1072366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1072411Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1072468Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1072567Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1072883Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1072959Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1073165Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1073254Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1073462Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1073550Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1073622Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1073665Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1073720Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1073818Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1074139Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1074213Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1074435Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1074532Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1074740Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1074828Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1074901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1074942Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1075001Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1075101Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1075420Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1075494Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1075703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1075789Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1075998Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1076102Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1076191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1076235Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1076292Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1076390Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1076709Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1076836Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1077045Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1077134Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1077340Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1077431Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1077502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1077547Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1077603Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1077703Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1078019Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1078115Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1078323Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1078427Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1078634Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1078722Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1078794Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1078838Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1078894Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1078994Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1079313Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1079388Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1079597Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1079683Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1079891Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1080015Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1080092Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1080134Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1080191Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1080289Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1080606Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1080679Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1080890Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1080977Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1081184Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1081271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1081344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1081385Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1081441Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1081539Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1081857Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1081949Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1082167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1082255Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1082461Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1082548Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1082620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1082663Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1082720Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1082820Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1083141Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1083216Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1083423Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1083510Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1083716Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1083824Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1083899Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1083942Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1083999Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1084097Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1084416Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1084488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1084698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1084786Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1084994Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1085080Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1085153Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1085195Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1085251Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1085348Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1085667Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1085759Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1085977Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1086064Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1086271Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1086358Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1086431Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1086474Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1086530Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1086630Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1087006Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1087080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1087290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1087378Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1087600Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1087704Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1087778Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1087822Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1087878Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1087977Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1088296Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1088369Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1088578Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1088666Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1088875Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1088962Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1089035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1089076Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1089133Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1089231Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1089552Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1092708Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1101891Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1101978Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1102188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1102273Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1102347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1102388Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1102446Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1102546Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1102863Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1102937Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1103148Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1103235Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1103463Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1103566Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1103639Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1103682Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1103737Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1103836Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1104151Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1104225Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1104433Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1104521Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1104728Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1104815Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1104886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1104928Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1104983Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1105083Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1105403Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1105495Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1105716Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1105802Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1106009Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1106094Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1106167Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1106207Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1106265Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1106363Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1106680Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1106799Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1107011Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1107098Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1107334Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1107422Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1107505Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1107551Z Traceback (most recent call last): 2025-12-04T09:54:35.1107674Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1107713Z return value(self) 2025-12-04T09:54:35.1107864Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1107912Z self.assertExpectedInline( 2025-12-04T09:54:35.1108089Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1108226Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1108377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1108423Z assert_expected_inline( 2025-12-04T09:54:35.1108570Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1108623Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1108788Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1108865Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1108982Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1109044Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1109137Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1109188Z raise self.failureException(msg) 2025-12-04T09:54:35.1109295Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1109429Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1109707Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1109848Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1109946Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1110045Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1110222Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1110228Z 2025-12-04T09:54:35.1110303Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1110452Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1110454Z 2025-12-04T09:54:35.1110541Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1110616Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1110658Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1110716Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1110827Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1111159Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1111235Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1111451Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1111540Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1111752Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1111840Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1111915Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1111957Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1112015Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1112113Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1112430Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1112503Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1112713Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1112803Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1113024Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1113123Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1113195Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1113239Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1113295Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1113395Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1113710Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1113786Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1113995Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1114085Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1114289Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1114376Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1114448Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1114491Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1114546Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1114657Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1114987Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1115064Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1115273Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1115361Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1115569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1115657Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1115732Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1115774Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1115833Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1115931Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1116256Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1116329Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1116538Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1116626Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1116895Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1117000Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1117075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1117116Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1117173Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1117272Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1117593Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1117668Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1117877Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1117966Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1118171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1118260Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1118332Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1118374Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1118430Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1118542Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1118874Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1118949Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1119157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1119245Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1119451Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1119540Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1119613Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1119658Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1119715Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1119816Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1120133Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1120207Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1120415Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1120502Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1120729Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1120828Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1120903Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1120946Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1121004Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1121101Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1121420Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1121494Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1121705Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1121793Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1122002Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1122088Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1122161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1122202Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1122260Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1122382Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1122702Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1122778Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1122986Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1123074Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1123281Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1123370Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1123444Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1123488Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1123543Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1123642Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1123957Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1124031Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1124244Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1124333Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1124553Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1124650Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1124725Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1124767Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1124825Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1124923Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1125244Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1125318Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1125529Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1125615Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1125824Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1125910Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1125984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1126025Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1126082Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1126199Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1126517Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1126592Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1126834Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1126919Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1127125Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1127212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1127286Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1127329Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1127384Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1127482Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1127799Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1127874Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1128084Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1128189Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1128394Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1128497Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1128568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1128612Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1128667Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1128767Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1129087Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1129162Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1129372Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1129460Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1129668Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1129754Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1129827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1129869Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1129941Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1130062Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1130382Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1130455Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1130665Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1130751Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1130959Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1131046Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1131121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1131164Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1131220Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1131318Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1131637Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1131711Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1131923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1132026Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1132243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1132331Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1132402Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1132445Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1132501Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1132601Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1132918Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1132993Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1133201Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1133288Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1133496Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1133583Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1133654Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1133698Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1133763Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1133871Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1134190Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1134265Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1134476Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1134562Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1134773Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1134863Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1134940Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1134984Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1135044Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1135143Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1135462Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1135536Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1135751Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1135852Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1136073Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1136161Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1136238Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1136281Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1136341Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1136441Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1136810Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1136888Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1137099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1137191Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1137398Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1137489Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1137571Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1137621Z Traceback (most recent call last): 2025-12-04T09:54:35.1137780Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1137826Z return value(self) 2025-12-04T09:54:35.1137978Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1138030Z self.assertExpectedInline( 2025-12-04T09:54:35.1138205Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1138345Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1138493Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1138541Z assert_expected_inline( 2025-12-04T09:54:35.1138689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1138743Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1138911Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1138993Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1139110Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1139177Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1139273Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1139327Z raise self.failureException(msg) 2025-12-04T09:54:35.1139417Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1139542Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1139820Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1139990Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1140092Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1140190Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1140370Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1140372Z 2025-12-04T09:54:35.1140450Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1140604Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1140608Z 2025-12-04T09:54:35.1140698Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1140775Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1140818Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1140880Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1140980Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1141303Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1141389Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1141614Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1141710Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1141920Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1142011Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1142085Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1142131Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1142189Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1142295Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1142617Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1142698Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1142907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1142997Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1143207Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1143303Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1143378Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1143436Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1143503Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1143602Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1143921Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1143998Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1144209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1144297Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1144510Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1144601Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1144680Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1144725Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1144787Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1144887Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1145207Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1145308Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1145521Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1145609Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1145821Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1145909Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1145987Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1146030Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1146090Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1146190Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1146515Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1146592Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1146879Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1146970Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1147176Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1147268Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1147341Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1147406Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1147479Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1147581Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1147900Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1147977Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1148188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1148280Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1148489Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1148583Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1148655Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1148702Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1148759Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1148859Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1149180Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1149280Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1149494Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1149584Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1149793Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1149881Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1149957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1149999Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1150062Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1150164Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1150491Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1150569Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1150780Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1150868Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1151079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1151168Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1151246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1151301Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1151373Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1151472Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1151792Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1151869Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1152077Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1152170Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1152380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1152473Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1152547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1152592Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1152648Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1152750Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1153080Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1153172Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1153382Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1153473Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1153682Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1153775Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1153852Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1153895Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1153955Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1154055Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1154375Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1154450Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1154663Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1154750Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1154960Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1155048Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1155135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1155190Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1155250Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1155349Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1155668Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1155742Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1155954Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1156043Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1156250Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1156338Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1156409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1156454Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1156510Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1156608Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1156999Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1157087Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1157296Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1157384Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1157588Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1157676Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1157748Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1157794Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1157850Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1157950Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1158266Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1158343Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1158551Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1158639Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1158847Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1158937Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1159024Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1159086Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1159144Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1159242Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1159560Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1159633Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1159845Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1159932Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1160141Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1160228Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1160302Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1160345Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1160403Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1160501Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1160841Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1160915Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1161124Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1161213Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1161425Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1161515Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1161587Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1161629Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1161685Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1161784Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1162103Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1162180Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1162387Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1162476Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1162682Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1162770Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1162852Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1162905Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1162961Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1163059Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1163376Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1163452Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1163663Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1163751Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1163960Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1164047Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1164122Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1164164Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1164223Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1164321Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1164656Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1164729Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1164939Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1165026Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1165232Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1165319Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1165392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1165433Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1165493Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1165592Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1165911Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1165985Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1166194Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1166283Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1166492Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1166580Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1166667Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1166723Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1166811Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1166910Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1167226Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1167300Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1167509Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1167598Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1167806Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1167895Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1167968Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1168011Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1168066Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1168165Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1168514Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1168590Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1168801Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1168888Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1169096Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1169183Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1169267Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1169314Z Traceback (most recent call last): 2025-12-04T09:54:35.1169436Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1169475Z return value(self) 2025-12-04T09:54:35.1169628Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1169676Z self.assertExpectedInline( 2025-12-04T09:54:35.1169853Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1169992Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1170142Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1170184Z assert_expected_inline( 2025-12-04T09:54:35.1170334Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1170385Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1170568Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1170656Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1170776Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1170838Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1170935Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1170983Z raise self.failureException(msg) 2025-12-04T09:54:35.1171074Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1171193Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1171471Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1171608Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1171708Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1171806Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1171982Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1171984Z 2025-12-04T09:54:35.1172071Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1172229Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1172234Z 2025-12-04T09:54:35.1172324Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1172397Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1172441Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1172497Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1172599Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1172921Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1172998Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1173214Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1173303Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1173513Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1173600Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1173676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1173718Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1173778Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1173875Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1174195Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1174292Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1174501Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1174589Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1174797Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1174883Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1174959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1175002Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1175059Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1175159Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1175477Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1175553Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1175761Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1175863Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1176078Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1176168Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1176240Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1176284Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1176340Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1176442Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1176798Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1176875Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1177086Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1177178Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1177383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1177471Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1177544Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1177589Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1177646Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1177748Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1178066Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1178167Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1178377Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1178463Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1178673Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1178759Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1178835Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1178877Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1178936Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1179035Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1179354Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1179427Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1179637Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1179737Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1179958Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1180047Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1180121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1180162Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1180219Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1180318Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1180639Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1180715Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1180923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1181012Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1181218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1181305Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1181378Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1181422Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1181478Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1181579Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1181906Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1181991Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1182199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1182286Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1182492Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1182581Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1182655Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1182697Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1182755Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1182854Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1183173Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1183245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1183456Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1183561Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1183775Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1183861Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1183934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1183976Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1184033Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1184132Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1184450Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1184524Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1184734Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1184823Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1185030Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1185119Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1185193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1185238Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1185293Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1185395Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1185724Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1185811Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1186021Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1186108Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1186315Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1186404Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1186478Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1186522Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1186578Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1190684Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1191017Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1191093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1191312Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1191459Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1191674Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1191764Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1191839Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1191881Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1191939Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1192038Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1192362Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1192439Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1192648Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1192737Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1192946Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1193033Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1193105Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1193148Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1193204Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1193305Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1193639Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1193730Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1193939Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1194027Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1194235Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1194324Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1194397Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1194441Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1194497Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1194596Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1194912Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1194986Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1195204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1195302Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1195512Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1195599Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1195672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1195714Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1195770Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1195868Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1196186Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1196260Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1196472Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1196559Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1196813Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1196900Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1196974Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1197015Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1197074Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1197172Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1197515Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1197603Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1197812Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1197899Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1198105Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1198194Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1198267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1198310Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1198366Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1198465Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1198784Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1198859Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1199083Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1199184Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1199392Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1199481Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1199553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1199596Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1199651Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1199750Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1200068Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1200142Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1200352Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1200439Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1200650Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1200735Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1200809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1200850Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1200908Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1201007Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1201345Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1201417Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1201626Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1201712Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1201923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1202009Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1202083Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1202125Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1202181Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1202277Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1202593Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1202666Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1202884Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1202984Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1203193Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1203280Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1203352Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1203394Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1203450Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1203547Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1203864Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1203938Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1204146Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1204234Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1204439Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1204525Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1204597Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1204639Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1204695Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1204807Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1205140Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1205213Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1205423Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1205508Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1205715Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1205803Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1205886Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1205933Z Traceback (most recent call last): 2025-12-04T09:54:35.1206056Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1206095Z return value(self) 2025-12-04T09:54:35.1206248Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1206298Z self.assertExpectedInline( 2025-12-04T09:54:35.1206474Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1206610Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1206828Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1206872Z assert_expected_inline( 2025-12-04T09:54:35.1207022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1207075Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1207240Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1207316Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1207435Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1207496Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1207591Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1207640Z raise self.failureException(msg) 2025-12-04T09:54:35.1207733Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1207857Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1208137Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1208276Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1208376Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1208477Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1208671Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1208689Z 2025-12-04T09:54:35.1208766Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1208914Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1208916Z 2025-12-04T09:54:35.1209005Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1209080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1209123Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1209179Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1209280Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1209604Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1209681Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1209895Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1209983Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1210192Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1210278Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1210363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1210417Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1210475Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1210574Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1210890Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1210962Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1211171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1211259Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1211466Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1211554Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1211626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1211668Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1211725Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1211824Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1212142Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1212216Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1212438Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1212538Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1212743Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1212830Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1212902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1212944Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1213000Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1213100Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1213416Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1213491Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1213698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1213786Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1213992Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1214080Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1214171Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1214215Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1214272Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1214371Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1214687Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1214759Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1214966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1215053Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1215261Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1215348Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1215420Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1215461Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1215517Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1215615Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1215931Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1216006Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1216230Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1216326Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1216533Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1216619Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1216692Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1216734Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1216833Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1216932Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1217251Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1217325Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1217533Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1217621Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1217827Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1217930Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1218017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1218061Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1218117Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1218216Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1218535Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1218608Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1218815Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1218903Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1219109Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1219197Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1219269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1219310Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1219367Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1219464Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1219780Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1219853Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1220074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1220179Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1220385Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1220471Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1220544Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1220586Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1220642Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1220741Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1221058Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1221131Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1221338Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1221424Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1221631Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1221729Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1221813Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1221857Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1221915Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1222014Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1222330Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1222404Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1222614Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1222704Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1222912Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1223001Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1223075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1223119Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1223176Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1223276Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1223595Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1223670Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1223890Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1223988Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1224196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1224283Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1224357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1224399Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1224457Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1224557Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1224877Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1224953Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1225165Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1225251Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1225460Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1225555Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1225639Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1225682Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1225740Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1225837Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1226155Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1226227Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1226438Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1226527Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1226734Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1226866Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1226939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1226982Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1227039Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1227139Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1227457Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1227546Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1227768Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1227856Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1228067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1228155Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1228228Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1228271Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1228327Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1228430Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1228746Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1228823Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1229033Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1229120Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1229331Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1229443Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1229520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1229563Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1229620Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1229717Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1230035Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1230107Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1230318Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1230406Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1230616Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1230702Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1230774Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1230816Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1230874Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1230971Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1231290Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1231381Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1231604Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1231693Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1231904Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1231992Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1232064Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1232108Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1232165Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1232264Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1232581Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1232657Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1232863Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1232953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1233168Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1233266Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1233340Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1233385Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1233441Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1233540Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1233859Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1233932Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1234140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1234227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1234435Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1234521Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1234594Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1234636Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1234693Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1234790Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1235108Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1235197Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1235416Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1235502Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1235709Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1235795Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1235870Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1235911Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1235969Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1236068Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1236386Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1236460Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1236667Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1236795Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1237016Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1237119Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1237192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1237236Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1237292Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1237391Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1237708Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1237782Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1237993Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1238083Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1238290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1238377Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1238449Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1238491Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1238547Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1238645Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1238966Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1239052Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1239274Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1239361Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1239569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1239655Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1239737Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1239784Z Traceback (most recent call last): 2025-12-04T09:54:35.1239908Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1239948Z return value(self) 2025-12-04T09:54:35.1240100Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1240149Z self.assertExpectedInline( 2025-12-04T09:54:35.1240327Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1240464Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1240612Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1240655Z assert_expected_inline( 2025-12-04T09:54:35.1240815Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1240877Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1241044Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1241119Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1241237Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1241297Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1241392Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1241440Z raise self.failureException(msg) 2025-12-04T09:54:35.1241531Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1241651Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1241926Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1242065Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1242165Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1242263Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1242443Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1242446Z 2025-12-04T09:54:35.1242524Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1242675Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1242695Z 2025-12-04T09:54:35.1242785Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1242860Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1242904Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1242961Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1243061Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1243381Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1243455Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1243669Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1243758Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1243966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1244053Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1244127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1244169Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1244226Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1244338Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1244671Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1244747Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1244956Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1245042Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1245250Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1245335Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1245410Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1245453Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1245509Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1245607Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1245923Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1245997Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1246203Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1246291Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1246499Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1246606Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1246678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1246721Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1246823Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1246921Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1247236Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1247311Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1247519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1247607Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1247811Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1247898Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1247970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1248011Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1248067Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1248178Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1248513Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1248588Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1248796Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1248883Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1251116Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1251205Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1251279Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1251322Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1251379Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1251476Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1251793Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1251866Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1252075Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1252163Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1252394Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1252496Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1252568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1252610Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1252665Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1252763Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1253083Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1253158Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1253368Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1253456Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1253662Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1253748Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1253820Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1253862Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1253918Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1254028Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1254357Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1254434Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1254642Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1254730Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1254936Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1255022Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1255095Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1255137Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1255194Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1255292Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1255608Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1255681Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1255889Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1255977Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1256199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1256297Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1256373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1256415Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1256473Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1256571Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1256938Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1257012Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1257226Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1257316Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1257526Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1257616Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1257687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1257731Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1257787Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1257917Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1258235Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1258310Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1258518Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1258611Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1258818Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1258909Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1258985Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1259030Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1259086Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1259188Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1259505Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1259580Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1259793Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1259882Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1260107Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1260207Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1260281Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1260324Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1260382Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1260481Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1260802Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1260880Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1261092Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1261179Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1261388Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1261474Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1261550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1261592Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1261651Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1261774Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1262099Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1262175Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1262384Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1262472Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1262679Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1262769Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1262843Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1262889Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1262944Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1263046Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1263362Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1263435Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1263644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1263735Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1263960Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1264059Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1264133Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1264179Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1264236Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1264337Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1264657Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1264731Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1264942Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1265030Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1265238Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1265324Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1265399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1265440Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1265516Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1265626Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1265948Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1266022Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1266232Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1266320Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1266529Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1266617Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1266695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1266775Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1266834Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1266932Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1267253Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1267327Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1267536Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1267645Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1267869Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1267958Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1268030Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1268074Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1268130Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1268230Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1268550Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1268627Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1268837Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1268927Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1269134Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1269222Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1269295Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1269338Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1269412Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1269524Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1269845Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1269919Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1270133Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1270220Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1270431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1270520Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1270597Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1270641Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1270699Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1270799Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1271124Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1271198Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1271411Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1271512Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1271734Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1271823Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1271897Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1271940Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1271996Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1272097Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1272416Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1272493Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1272703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1272792Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1272999Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1273088Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1273161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1273208Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1273277Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1273390Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1273710Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1273785Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1273990Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1274080Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1274289Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1274378Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1274453Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1274496Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1274554Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1274653Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1274971Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1275046Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1275259Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1275359Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1275581Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1275667Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1275743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1275785Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1275845Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1275943Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1276263Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1276337Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1276549Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1276639Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1276893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1276982Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1277064Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1277126Z Traceback (most recent call last): 2025-12-04T09:54:35.1277260Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1277303Z return value(self) 2025-12-04T09:54:35.1277459Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1277512Z self.assertExpectedInline( 2025-12-04T09:54:35.1277689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1277829Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1277980Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1278027Z assert_expected_inline( 2025-12-04T09:54:35.1278174Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1278229Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1278394Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1278474Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1278592Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1278655Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1278749Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1278803Z raise self.failureException(msg) 2025-12-04T09:54:35.1278896Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1279018Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1279296Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1279471Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1279574Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1279674Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1279854Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1279856Z 2025-12-04T09:54:35.1279933Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1280085Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1280088Z 2025-12-04T09:54:35.1280176Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1280252Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1280298Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1280358Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1280460Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1280785Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1280871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1281100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1281191Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1281402Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1281493Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1281567Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1281612Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1281672Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1281777Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1282097Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1282175Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1282385Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1282476Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1282684Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1282773Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1282846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1282905Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1282976Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1283075Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1283398Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1283475Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1283684Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1283776Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1283991Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1284081Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1284159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1284202Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1284261Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1284362Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1284683Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1284779Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1284993Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1285083Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1285296Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1285383Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1285459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1285502Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1285559Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1285659Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1285980Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1286054Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1286268Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1286358Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1286566Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1286658Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1286733Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1286830Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1286903Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1287002Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1287318Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1287393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1287601Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1287691Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1287901Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1287993Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1288066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1288112Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1288169Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1288269Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1288603Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1288694Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1288908Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1288995Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1289205Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1289292Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1289366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1289408Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1289466Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1289567Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1289890Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1289965Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1290179Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1290266Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1290477Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1290564Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1290653Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1290706Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1290763Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1290863Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1291183Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1291257Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1291464Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1291555Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1291761Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1291849Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1291921Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1291964Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1292020Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1292118Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1292445Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1292531Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1292737Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1292827Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1293034Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1293122Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1293193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1293237Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1293293Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1293397Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1293715Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1293790Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1294000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1294086Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1294295Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1294382Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1294476Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1294528Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1294585Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1294684Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1295003Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1295076Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1295287Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1295374Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1295584Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1295672Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1295747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1295789Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1295848Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1295947Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1296278Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1296365Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1296574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1296662Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1296905Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1296994Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1297066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1297109Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1297165Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1297265Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1297583Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1297659Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1297865Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1297954Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1298160Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1298249Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1298338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1298394Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1298451Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1298550Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1298866Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1298940Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1299150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1299237Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1299447Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1299535Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1299608Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1299649Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1299707Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1299804Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1300146Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1300220Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1300430Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1300517Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1300727Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1300816Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1300888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1300931Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1300988Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1301088Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1301408Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1301483Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1301693Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1301781Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1301989Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1302078Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1302163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1302216Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1302272Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1302370Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1302685Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1302759Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1302966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1303056Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1303263Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1303349Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1303423Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1303464Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1303520Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1303618Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1303960Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1304034Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1304243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1304329Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1304536Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1304622Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1304696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1304739Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1304797Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1304896Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1305214Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1305287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1305496Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1305583Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1305791Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1305892Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1305976Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1306019Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1306074Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1306174Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1306490Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1306564Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1306812Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1306901Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1307109Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1307196Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1307267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1307310Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1307365Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1307462Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1307812Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1307887Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1308100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1308186Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1308395Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1308481Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1308555Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1308597Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1308657Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1308755Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1309074Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1309147Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1309356Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1309442Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1309651Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1309753Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1309838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1309880Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1309936Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1310033Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1310356Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1310430Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1310641Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1310731Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1310938Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1311025Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1311097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1311140Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1311195Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1311294Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1311629Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1311706Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1311915Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1312004Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1312210Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1312296Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1312369Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1312412Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1312469Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1312570Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1312890Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1312962Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1313173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1313259Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1313468Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1315710Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1319701Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1319748Z Traceback (most recent call last): 2025-12-04T09:54:35.1319869Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1319908Z return value(self) 2025-12-04T09:54:35.1320060Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1320108Z self.assertExpectedInline( 2025-12-04T09:54:35.1320284Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1320421Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1320572Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1320615Z assert_expected_inline( 2025-12-04T09:54:35.1320764Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1320816Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1320982Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1321059Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1321178Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1321239Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1328866Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1328932Z raise self.failureException(msg) 2025-12-04T09:54:35.1329029Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1329151Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1329429Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1329565Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1329664Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1329764Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1329942Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1329945Z 2025-12-04T09:54:35.1330022Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1330168Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1330171Z 2025-12-04T09:54:35.1330259Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1330336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1330380Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1330436Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1330541Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1330881Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1330977Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1331189Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1331276Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1331484Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1331572Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1331647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1331690Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1331748Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1331848Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1332164Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1332236Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1332446Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1332547Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1332775Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1332863Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1332936Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1332977Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1333034Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1333132Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1333451Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1333525Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1333735Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1333824Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1334030Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1334117Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1334188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1334232Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1334288Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1334389Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1334719Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1334807Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1335017Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1335105Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1335313Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1335402Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1335476Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1335518Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1335576Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1335675Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1335995Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1336069Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1336283Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1336405Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1336614Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1336703Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1336797Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1336839Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1336896Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1336995Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1337320Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1337393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1337602Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1337691Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1337898Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1337984Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1338057Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1338098Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1338155Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1338254Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1338586Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1338674Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1338881Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1338969Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1339177Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1339265Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1339338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1339382Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1339438Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1339537Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1339854Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1339929Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1340152Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1340253Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1340461Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1340548Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1340622Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1340665Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1340723Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1340823Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1341142Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1341218Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1341430Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1341516Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1341724Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1341810Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1341883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1341925Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1341981Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1342083Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1342415Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1342500Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1342708Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1342794Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1343001Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1343090Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1343163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1343206Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1343261Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1343360Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1343681Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1343754Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1343972Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1344071Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1344277Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1344367Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1344439Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1344482Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1344537Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1344635Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1344953Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1345031Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1345239Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1345327Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1345534Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1345621Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1345695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1345737Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1345796Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1345894Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1346225Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1346309Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1346520Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1346608Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1346864Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1346953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1347027Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1347070Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1347128Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1347228Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1347547Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1347620Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1347850Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1347956Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1348168Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1348256Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1348328Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1348371Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1348429Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1348530Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1348847Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1348923Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1349132Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1349221Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1349427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1349518Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1349590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1349633Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1349690Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1349806Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1350143Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1350217Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1350429Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1350516Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1350726Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1350814Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1350888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1350931Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1350988Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1351087Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1351405Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1351477Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1351706Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1351798Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1352006Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1352093Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1352167Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1352209Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1352267Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1352368Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1352691Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1352766Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1352975Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1353064Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1353270Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1353360Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1353432Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1353475Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1353532Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1353646Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1353974Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1354048Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1354256Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1354345Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1354553Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1354643Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1354718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1354762Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1354817Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1354917Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1355234Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1355307Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1355546Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1355635Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1355844Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1355930Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1356004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1356045Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1356103Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1356200Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1356518Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1356592Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1356835Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1356923Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1357130Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1357216Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1357290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1357333Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1357392Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1357508Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1357838Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1357912Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1358118Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1358207Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1358415Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1358505Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1358578Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1358622Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1358678Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1358778Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1359098Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1359172Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1359407Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1359498Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1359705Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1359794Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1359868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1359910Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1359968Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1360066Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1360387Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1360461Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1360671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1360758Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1360966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1361054Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1361128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1361171Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1361246Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1361344Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1361673Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1361746Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1361955Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1362044Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1362254Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1362343Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1362416Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1362460Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1362515Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1362615Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1362931Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1363005Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1363232Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1363324Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1363532Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1363621Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1363693Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1363737Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1363793Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1363893Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1364212Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1364287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1364497Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1364586Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1364795Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1364882Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1364966Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1365013Z Traceback (most recent call last): 2025-12-04T09:54:35.1365154Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1365207Z return value(self) 2025-12-04T09:54:35.1365360Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1365410Z self.assertExpectedInline( 2025-12-04T09:54:35.1365590Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1365727Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1365877Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1365920Z assert_expected_inline( 2025-12-04T09:54:35.1366069Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1366121Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1366288Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1366364Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1366483Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1366545Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1366639Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1366687Z raise self.failureException(msg) 2025-12-04T09:54:35.1366814Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1366951Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1367242Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1367380Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1367481Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1367581Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1367760Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1367762Z 2025-12-04T09:54:35.1367840Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1367989Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1367993Z 2025-12-04T09:54:35.1368083Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1368156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1368201Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1368257Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1368360Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1368680Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1368756Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1368984Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1369091Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1369298Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1369388Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1369463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1369505Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1369563Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1369664Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1369988Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1370062Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1370271Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1370359Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1370568Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1370668Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1370756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1370799Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1370858Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1370956Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1371274Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1371347Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1371557Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1371645Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1371853Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1371945Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1372019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1372063Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1372119Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1372223Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1372540Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1372628Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1372845Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1372933Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1373139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1373228Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1373301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1373346Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1373402Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1373503Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1373820Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1373894Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1374102Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1374193Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1374401Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1374508Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1374584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1374628Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1374685Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1374783Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1375101Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1375174Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1375385Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1375472Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1375680Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1375767Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1377926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1377971Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1378029Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1378127Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1378450Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1378553Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1378779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1378867Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1379076Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1379163Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1379236Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1379278Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1379336Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1379436Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1379755Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1379832Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1380040Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1380127Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1380334Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1380457Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1380533Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1380580Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1380636Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1380734Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1381050Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1381123Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1381331Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1381422Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1381631Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1381716Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1381788Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1381830Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1381887Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1381984Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1382302Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1382388Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1382614Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1382700Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1382907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1382994Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1383067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1383109Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1383166Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1383266Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1383584Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1383657Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1383863Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1383951Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1384166Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1384263Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1384337Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1384380Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1384436Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1384535Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1384857Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1384931Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1385140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1385228Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1385436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1385523Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1385595Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1385637Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1385693Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1385792Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1386112Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1386202Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1386422Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1386509Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1386716Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1386835Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1386910Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1386951Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1387009Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1387107Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1387426Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1387498Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1387706Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1387792Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1388014Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1388116Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1388191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1388232Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1388289Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1388386Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1388706Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1388779Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1388988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1389076Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1389289Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1389375Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1389447Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1389489Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1389544Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1389643Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1389960Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1390050Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1390270Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1390357Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1390563Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1390649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1390721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1390764Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1390820Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1390919Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1391240Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1391313Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1391522Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1391608Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1391836Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1391923Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1391997Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1392039Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1392096Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1392193Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1392512Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1392585Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1392796Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1392884Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1393095Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1393182Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1393254Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1393296Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1393352Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1393452Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1393771Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1393873Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1394080Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1394167Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1394373Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1394460Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1394532Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1394575Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1394633Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1394732Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1395049Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1395123Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1395330Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1395417Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1395655Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1395743Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1395817Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1395859Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1395915Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1396012Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1396329Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1396401Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1396610Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1396700Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1396951Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1397038Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1397112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1397153Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1397210Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1397307Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1397627Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1397734Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1397944Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1398031Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1398237Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1398324Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1398396Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1398438Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1398496Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1398594Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1398910Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1398983Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1399190Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1399277Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1399508Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1399597Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1399669Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1399712Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1399767Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1399866Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1400185Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1400258Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1400468Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1400555Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1400762Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1400848Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1400920Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1400961Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1401017Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1401115Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1401433Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1401526Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1401734Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1401820Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1402027Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1402113Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1402188Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1402230Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1402288Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1402388Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1402706Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1402780Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1402987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1403073Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1403299Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1403389Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1403460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1403503Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1403558Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1403656Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1403971Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1404045Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1404254Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1404343Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1404550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1404637Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1404709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1404752Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1404807Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1404904Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1405221Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1405317Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1405526Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1405611Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1405819Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1405904Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1405988Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1406036Z Traceback (most recent call last): 2025-12-04T09:54:35.1406158Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1406198Z return value(self) 2025-12-04T09:54:35.1406350Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1406399Z self.assertExpectedInline( 2025-12-04T09:54:35.1406577Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1406713Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1406894Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1406937Z assert_expected_inline( 2025-12-04T09:54:35.1407111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1407163Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1407329Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1407405Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1407524Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1407585Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1407679Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1407728Z raise self.failureException(msg) 2025-12-04T09:54:35.1407822Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1407945Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1408224Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1408363Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1408462Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1408561Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1408749Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1408752Z 2025-12-04T09:54:35.1408848Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1408997Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1409021Z 2025-12-04T09:54:35.1409112Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1409186Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1409231Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1409287Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1409387Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1409709Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1409785Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1410000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1410089Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1410300Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1410387Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1410461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1410503Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1410561Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1410682Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1411006Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1411080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1411289Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1411377Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1411587Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1411676Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1411751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1411793Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1411851Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1411949Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1412268Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1412341Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1412550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1412654Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1412859Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1412959Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1413035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1413077Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1413133Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1413232Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1413550Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1413624Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1413834Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1413922Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1414128Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1414216Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1414288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1414330Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1414397Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1414506Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1414824Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1414896Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1415105Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1415192Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1415401Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1415489Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1415563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1415609Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1415665Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1415763Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1416079Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1416152Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1416361Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1416458Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1416679Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1416798Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1416871Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1416913Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1416969Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1417067Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1417388Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1417462Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1417671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1417759Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1417965Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1418052Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1418123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1418166Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1418236Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1418349Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1418670Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1418743Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1418951Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1419037Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1419243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1419332Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1419404Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1419449Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1419505Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1419603Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1419922Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1419993Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1420203Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1420302Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1420525Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1420612Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1420685Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1420726Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1420782Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1420879Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1421198Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1421271Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1421480Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1421566Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1421774Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1421861Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1421933Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1421975Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1422049Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1422149Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1422465Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1422538Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1422745Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1422831Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1423038Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1423126Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1423199Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1423241Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1423296Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1423394Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1423709Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1423783Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1423995Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1424100Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1424317Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1424403Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1424476Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1424517Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1424573Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1424671Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1424990Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1425064Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1425272Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1425358Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1425567Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1425653Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1425725Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1425779Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1425846Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1425945Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1426266Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1426339Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1426548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1426635Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1426895Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1426983Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1427056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1427098Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1427153Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1427251Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1427567Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1427640Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1427848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1427951Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1428169Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1428255Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1428329Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1428371Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1428426Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1428524Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1428841Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1428916Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1429125Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1429211Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1429418Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1429504Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1429576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1429630Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1429701Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1429799Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1430115Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1430187Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1430395Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1430481Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1430694Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1430781Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1430854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1430895Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1430951Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1431048Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1431363Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1431437Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1431644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1431751Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1431958Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1432046Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1432118Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1432161Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1432216Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1432316Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1432639Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1432715Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1432923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1433011Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1433217Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1433304Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1433386Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1433439Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1433496Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1433597Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1433914Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1433987Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1434197Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1434283Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1434490Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1434577Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1434651Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1434692Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1434748Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1434846Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1435163Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1435236Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1435445Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1435551Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1435759Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1435845Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1435918Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1435960Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1436017Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1436114Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1436433Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1436508Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1436717Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1436842Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1437049Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1437137Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1437227Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1437290Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1437346Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1437445Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1437761Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1437835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1438042Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1438129Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1438336Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1438427Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1438499Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1438540Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1438597Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1438695Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1439011Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1439085Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1439295Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1439411Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1439617Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1439702Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1439776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1439818Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1439874Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1439971Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1440289Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1440366Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1440574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1440660Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1440866Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1440953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1441036Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1441090Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1441147Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1441249Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1441569Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1441642Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1441849Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1441937Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1442143Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1442233Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1442305Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1442347Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1442402Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1442500Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1442816Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1442891Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1443113Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1443215Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1443422Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1443508Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1443582Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1443624Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1443681Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1443780Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1444099Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1444175Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1444384Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1444470Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1444677Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1444763Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1444854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1444898Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1444956Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1445053Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1445371Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1445443Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1445651Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1445738Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1445947Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1446035Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1446116Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1446164Z Traceback (most recent call last): 2025-12-04T09:54:35.1446284Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1446323Z return value(self) 2025-12-04T09:54:35.1446473Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1446522Z self.assertExpectedInline( 2025-12-04T09:54:35.1446697Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1446874Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1447054Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1447098Z assert_expected_inline( 2025-12-04T09:54:35.1447245Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1447298Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1447461Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1447538Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1447655Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1447717Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1447812Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1447862Z raise self.failureException(msg) 2025-12-04T09:54:35.1447954Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1448076Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1448353Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1448491Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1448621Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1448720Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1448899Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1448901Z 2025-12-04T09:54:35.1448976Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1449126Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1449128Z 2025-12-04T09:54:35.1449215Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1449290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1449332Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1449390Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1449491Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1449814Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1449888Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1450100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1450188Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1450398Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1450498Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1450571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1450624Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1450680Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1450779Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1451094Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1451168Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1451377Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1451465Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1451674Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1451761Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1451835Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1451878Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1451934Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1452034Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1452378Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1452453Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1452662Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1452750Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1452956Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1453042Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1453115Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1453156Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1453214Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1453313Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1453631Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1453703Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1453911Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1453997Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1454206Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1454304Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1454387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1454429Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1454486Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1454583Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1454904Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1454978Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1455191Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1455279Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1455486Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1455574Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1455646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1455688Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1455744Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1455844Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1456187Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1456264Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1456470Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1456557Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1456810Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1456898Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1456970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1457013Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1457070Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1457170Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1457486Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1457560Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1457768Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1457854Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1458061Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1458162Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1458253Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1458295Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1458353Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1458451Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1458768Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1458839Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1459048Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1459135Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1459342Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1459428Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1459501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1459541Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1459597Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1459694Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1460039Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1460117Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1460323Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1460410Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1460615Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1460703Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1460774Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1460818Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1460876Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1460974Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1461294Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1461368Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1461577Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1461664Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1461871Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1461968Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1462050Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1462093Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1462151Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1462251Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1462570Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1462642Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1462852Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1462940Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1463152Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1463238Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1463313Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1463355Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1463414Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1463513Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1463855Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1463931Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1464142Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1464229Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1464438Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1464525Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1464599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1464642Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1464701Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1464797Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1465114Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1465190Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1465398Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1465488Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1465696Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1465800Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1465895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1465938Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1465995Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1466096Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1466418Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1466495Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1466705Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1466842Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1467047Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1467137Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1467213Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1467255Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1467311Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1467410Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1467760Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1467839Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1468050Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1468136Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1468347Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1468434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1468510Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1468552Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1468611Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1468711Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1469028Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1469102Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1469310Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1469396Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1469606Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1469726Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1469799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1469845Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1469902Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1470002Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1470324Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1470401Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1470609Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1470699Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1470908Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1470995Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1471068Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1471112Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1471169Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1471268Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1484265Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1484349Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1484556Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1484642Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1484848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1484934Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1485010Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1485053Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1485114Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1485214Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1485532Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1485606Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1485817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1485902Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1486115Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1486239Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1486313Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1486355Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1486411Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1486509Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1486870Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1486944Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1487158Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1487249Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1487458Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1487547Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1487619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1487664Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1487721Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1487819Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1488165Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1488241Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1488451Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1488539Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1488749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1488839Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1488913Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1488957Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1489013Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1489114Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1489434Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1489510Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1489721Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1489809Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1490019Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1490135Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1490208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1490249Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1490305Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1490404Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1490722Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1490796Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1491008Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1491095Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1491303Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1491390Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1491464Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1491505Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1491562Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1491675Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1492008Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1492085Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1492292Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1492380Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1492587Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1492675Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1492749Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1492792Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1492849Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1492948Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1493264Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1493337Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1493545Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1493634Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1493862Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1493961Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1494034Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1494076Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1494132Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1494231Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1494549Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1494622Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1494831Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1494918Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1495126Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1495212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1495285Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1495326Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1495382Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1495492Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1495824Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1495899Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1496109Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1496196Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1496405Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1496491Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1496566Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1496608Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1496665Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1496804Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1497121Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1497194Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1497400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1497489Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1497716Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1497818Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1497890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1497933Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1497988Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1498087Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1498412Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1498486Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1498695Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1498786Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1498993Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1499081Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1499154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1499195Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1499251Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1499373Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1499692Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1499765Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1499974Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1500060Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1500266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1500353Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1500436Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1500484Z Traceback (most recent call last): 2025-12-04T09:54:35.1500607Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1500646Z return value(self) 2025-12-04T09:54:35.1500798Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1500846Z self.assertExpectedInline( 2025-12-04T09:54:35.1501022Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1501159Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1501310Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1501365Z assert_expected_inline( 2025-12-04T09:54:35.1501514Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1501576Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1501741Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1501816Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1501935Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1501994Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1502088Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1502137Z raise self.failureException(msg) 2025-12-04T09:54:35.1502230Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1502353Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1502629Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1502766Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1502866Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1502963Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1503160Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1503163Z 2025-12-04T09:54:35.1503240Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1503388Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1503390Z 2025-12-04T09:54:35.1503479Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1503553Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1503597Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1503653Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1503754Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1504076Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1504151Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1504363Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1504451Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1504660Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1504747Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1504822Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1504864Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1504922Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1505035Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1505367Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1505440Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1505650Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1505738Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1505946Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1506034Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1506109Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1506151Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1506208Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1506306Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1506624Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1506698Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1506973Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1507063Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1507273Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1507359Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1507432Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1507474Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1507530Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1507628Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1507945Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1508021Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1508227Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1508314Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1508520Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1508607Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1508679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1508721Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1508796Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1508894Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1509230Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1509302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1509510Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1509596Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1509805Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1509891Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1509965Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1510006Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1510063Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1510160Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1510476Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1510548Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1510777Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1510866Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1511072Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1511158Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1511230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1511272Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1511328Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1511426Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1511744Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1511819Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1512026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1512112Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1512318Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1512405Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1512477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1512520Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1512588Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1512704Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1513020Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1513093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1513298Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1513385Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1513597Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1513683Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1513756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1513798Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1513855Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1513951Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1514267Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1514349Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1514567Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1514655Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1514862Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1514949Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1515021Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1515062Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1515118Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1515216Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1515535Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1515611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1515822Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1515908Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1516114Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1516201Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1516274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1516327Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1516383Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1516491Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1516849Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1516923Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1517130Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1517217Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1517427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1517517Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1517589Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1517632Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1517688Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1517787Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1518108Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1518198Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1518418Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1518507Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1518715Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1518801Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1518874Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1518915Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1518973Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1519070Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1519390Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1519464Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1519675Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1519761Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1519971Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1520058Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1520132Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1520188Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1520244Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1520354Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1520670Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1520744Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1520950Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1521037Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1521242Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1521331Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1521403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1521446Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1521501Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1521600Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1521915Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1522007Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1522231Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1522321Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1522526Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1522617Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1522688Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1522731Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1522787Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1522886Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1523207Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1523280Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1523488Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1523574Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1523781Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1523867Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1523941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1524000Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1524069Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1524167Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1524488Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1524561Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1524769Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1524856Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1525066Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1525153Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1525226Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1525268Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1525324Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1525421Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1525737Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1525829Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1526036Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1526125Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1526330Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1526417Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1526489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1526532Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1526588Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1526689Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1527045Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1527121Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1527329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1527418Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1527624Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1527712Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1527787Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1527849Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1527921Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1528018Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1528335Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1528407Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1528616Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1528704Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1528914Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1529000Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1529074Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1529115Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1529172Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1529270Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1529599Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1529684Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1529893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1529981Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1530188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1530276Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1530347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1530390Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1530446Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1530545Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1530862Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1530938Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1531147Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1531235Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1531440Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1531528Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1531618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1531670Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1531725Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1531823Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1532138Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1532211Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1532419Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1532508Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1532716Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1532802Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1532874Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1532915Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1532971Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1533069Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1533404Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1533486Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1533694Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1533781Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1533988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1534073Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1534146Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1534187Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1534244Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1534343Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1534659Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1534733Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1534941Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1535027Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1535234Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1535321Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1535406Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1535464Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1535518Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1535615Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1535933Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1536006Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1536213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1536301Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1536508Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1536597Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1536669Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1536712Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1536800Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1536898Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1537228Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1537315Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1537524Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1537610Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1537821Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1537906Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1537979Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1538022Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1538079Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1538177Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1538495Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1538569Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1538779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1538866Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1539074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1539163Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1539253Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1539309Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1539367Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1539465Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1539789Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1539866Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1540075Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1540164Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1540369Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1540458Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1540531Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1540576Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1540631Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1540733Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1541069Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1541146Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1541356Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1541446Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1541652Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1541741Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1541815Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1541857Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1541916Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1542018Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1542341Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1542414Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1542624Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1542711Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1542921Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1543007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1543103Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1543161Z Traceback (most recent call last): 2025-12-04T09:54:35.1543283Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1543325Z return value(self) 2025-12-04T09:54:35.1543478Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1543527Z self.assertExpectedInline( 2025-12-04T09:54:35.1543704Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1543841Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1543992Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1544036Z assert_expected_inline( 2025-12-04T09:54:35.1544185Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1544236Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1544402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1544480Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1544598Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1544658Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1544752Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1544811Z raise self.failureException(msg) 2025-12-04T09:54:35.1544915Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1545036Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1545312Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1545449Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1545547Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1545647Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1545826Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1545830Z 2025-12-04T09:54:35.1545910Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1546059Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1546061Z 2025-12-04T09:54:35.1546152Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1546229Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1546275Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1546333Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1546433Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1546795Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1546898Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1547114Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1547203Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1547413Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1547499Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1547575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1547619Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1547677Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1547777Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1548096Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1548168Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1548379Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1548467Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1548703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1548795Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1548870Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1548913Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1548972Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1549072Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1549392Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1549466Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1549675Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1549766Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1549972Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1550058Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1550131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1550176Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1550231Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1550331Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1550652Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1550761Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1550968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1551055Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1551259Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1551346Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1551420Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1551467Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1551526Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1551628Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1551946Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1552020Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1552227Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1552324Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1552543Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1552632Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1552706Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1552749Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1552805Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1552904Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1553226Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1553301Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1553511Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1553600Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1553808Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1553894Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1553970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1554013Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1554071Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1554171Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1554490Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1554588Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1554796Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1554886Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1555100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1555188Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1555263Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1555309Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1555366Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1555465Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1555782Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1555855Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1556071Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1556173Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1556392Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1556483Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1556560Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1556604Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1556662Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1556805Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1557124Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1557202Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1557414Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1557503Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1557713Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1557800Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1557875Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1557918Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1557975Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1558077Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1558417Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1558507Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1558717Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1558804Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1559015Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1559103Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1559175Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1559221Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1559279Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1559381Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1561543Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1561632Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1561840Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1561951Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1562177Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1562267Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1562338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1562381Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1562437Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1562536Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1562854Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1562930Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1563139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1563227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1563433Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1563520Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1563594Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1563635Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1563691Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1563790Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1564123Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1564209Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1564420Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1564506Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1564714Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1564801Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1564879Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1564920Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1564977Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1565074Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1565390Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1565463Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1565671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1565783Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1565992Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1566079Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1566151Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1566194Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1566250Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1566349Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1566666Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1566780Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1566989Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1567078Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1567284Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1567372Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1567443Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1567485Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1567541Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1567641Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1567978Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1568065Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1568271Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1568357Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1568563Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1568650Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1568724Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1568766Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1568823Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1568920Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1569236Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1569308Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1569531Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1569632Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1569841Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1569927Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1570000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1570041Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1570097Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1570195Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1570513Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1570588Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1570797Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1570884Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1571090Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1571176Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1571248Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1571290Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1571346Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1571445Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1571775Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1571861Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1572069Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1572156Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1572365Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1572453Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1572526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1572569Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1572625Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1572725Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1573047Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1573121Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1573340Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1573437Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1573644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1573731Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1573804Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1573845Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1573902Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1573999Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1574319Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1574393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1574603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1574689Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1574896Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1574982Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1575054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1575095Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1575152Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1575251Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1575578Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1575661Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1575868Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1575955Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1576162Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1576249Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1576322Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1576365Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1576420Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1576518Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1576872Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1576945Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1577169Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1577271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1577478Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1577565Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1577637Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1577679Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1577734Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1577833Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1578152Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1578225Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1578434Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1578522Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1578729Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1578815Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1578890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1578930Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1578988Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1579104Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1579432Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1579505Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1579713Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1579799Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1580006Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1580093Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1580166Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1580209Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1580265Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1580364Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1580681Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1580754Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1580987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1581077Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1581286Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1581372Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1581444Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1581486Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1581541Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1581640Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1581957Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1582031Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1582238Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1582325Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1582530Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1582616Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1582689Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1582730Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1582787Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1582896Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1583228Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1583300Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1583509Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1583595Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1583803Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1583891Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1583965Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1584006Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1584063Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1584160Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1584479Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1584552Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1584788Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1584876Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1585085Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1585175Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1585246Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1585290Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1585346Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1585445Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1585763Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1585837Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1586047Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1586135Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1586342Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1586429Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1586501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1586545Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1586601Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1586714Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1587083Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1587156Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1587362Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1587449Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1587663Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1587751Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1587824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1587865Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1587921Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1588018Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1588333Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1588405Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1588641Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1588729Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1588937Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1589023Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1589105Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1589152Z Traceback (most recent call last): 2025-12-04T09:54:35.1589274Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1589313Z return value(self) 2025-12-04T09:54:35.1589466Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1589517Z self.assertExpectedInline( 2025-12-04T09:54:35.1589695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1589832Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1589987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1590031Z assert_expected_inline( 2025-12-04T09:54:35.1590178Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1590230Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1590393Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1590470Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1590615Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1590689Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1590782Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1590830Z raise self.failureException(msg) 2025-12-04T09:54:35.1590921Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1591043Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1591317Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1591455Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1591556Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1591654Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1591833Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1591835Z 2025-12-04T09:54:35.1591911Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1592062Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1592065Z 2025-12-04T09:54:35.1592153Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1592248Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1592292Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1592351Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1592449Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1592770Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1592843Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1593056Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1593146Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1593356Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1593443Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1593516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1593557Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1593614Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1593712Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1594033Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1594107Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1594329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1594438Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1594644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1594731Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1594803Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1594846Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1594902Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1595002Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1595321Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1595395Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1595602Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1595689Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1595894Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1595998Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1596084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1596130Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1596186Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1596285Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1596604Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1596678Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1596933Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1597020Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1597228Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1597314Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1597387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1597429Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1597488Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1597585Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1597901Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1597975Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1598199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1598298Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1598507Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1598594Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1598667Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1598708Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1598764Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1598863Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1599181Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1599255Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1599461Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1599547Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1599752Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1599867Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1599953Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1599997Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1600052Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1600150Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1600466Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1600540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1600749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1600841Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1601047Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1601135Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1601207Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1601250Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1601304Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1601402Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1601722Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1601812Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1602020Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1602118Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1602327Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1602412Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1602485Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1602526Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1602582Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1602680Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1603007Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1603080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1603288Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1603373Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1603579Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1603686Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1603762Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1603805Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1603861Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1603959Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1604286Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1604359Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1604567Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1604655Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1604863Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1604950Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1605025Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1605067Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1605123Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1605222Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1605540Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1605628Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1605847Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1605934Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1606140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1606227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1606299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1606342Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1606398Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1606498Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1606857Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1606930Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1607139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1607225Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1607434Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1607552Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1607627Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1607669Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1607725Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1607822Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1608141Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1608214Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1608424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1608513Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1608721Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1608808Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1608880Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1608922Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1608977Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1609076Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1609397Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1609492Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1609713Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1609800Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1610008Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1610095Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1610166Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1610209Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1610265Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1610364Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1610682Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1610755Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1610963Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1611050Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1611270Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1611368Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1611442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1611484Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1611540Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1611640Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1611959Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1612031Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1612241Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1612328Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1612536Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1612621Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1612694Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1612736Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1612792Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1612888Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1613208Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1613295Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1613513Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1613601Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1613808Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1613895Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1613967Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1614010Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1614066Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1614166Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1614482Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1614556Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1614764Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1614851Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1615074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1615177Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1615250Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1615294Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1615349Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1615447Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1615764Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1615841Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1616054Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1616141Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1616349Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1616434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1616507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1616548Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1616604Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1616701Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1617141Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1617229Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1617454Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1617541Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1617748Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1617834Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1617909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1617955Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1618015Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1618113Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1618434Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1618507Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1618718Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1618805Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1619036Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1619127Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1619202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1619244Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1619300Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1619398Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1619714Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1619787Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1619997Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1620086Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1620295Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1620384Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1620459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1620505Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1620560Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1620658Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1620977Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1621082Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1621290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1621376Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1621583Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1621669Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1621742Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1621784Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1621841Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1621942Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1622259Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1622331Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1622538Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1622624Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1622864Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1622951Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1623025Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1623066Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1623123Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1623220Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1623543Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1623617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1623826Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1623915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1624125Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1624212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1624284Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1624326Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1624382Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1624480Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1624802Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1624898Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1625106Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1625195Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1625402Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1625488Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1625560Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1625604Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1625663Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1625762Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1626080Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1626152Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1626361Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1626446Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1626672Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1626795Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1626870Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1626912Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1626968Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1627067Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1627388Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1627461Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1627670Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1627758Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1627966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1628053Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1628126Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1628171Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1628227Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1628327Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1628644Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1628748Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1628954Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1629043Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1629254Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1629342Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1629416Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1629459Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1629515Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1629614Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1629930Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1630003Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1630209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1630296Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1630531Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1630619Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1630691Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1630734Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1630791Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1630888Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1631206Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1631280Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1631492Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1631581Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1631786Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1631871Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1631946Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1631990Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1632051Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1632148Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1632471Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1632570Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1632781Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1632868Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1633074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1633161Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1633233Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1633277Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1633333Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1633435Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1633757Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1633832Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1634040Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1634127Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1634355Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1634445Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1634526Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1634575Z Traceback (most recent call last): 2025-12-04T09:54:35.1634696Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1634737Z return value(self) 2025-12-04T09:54:35.1634887Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1634937Z self.assertExpectedInline( 2025-12-04T09:54:35.1635113Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1635253Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1635402Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1635447Z assert_expected_inline( 2025-12-04T09:54:35.1635596Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1635650Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1635818Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1635894Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1636014Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1636075Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1636174Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1636239Z raise self.failureException(msg) 2025-12-04T09:54:35.1636344Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1636464Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1636786Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1636925Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1637026Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1637124Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1637304Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1637307Z 2025-12-04T09:54:35.1637383Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1637534Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1637536Z 2025-12-04T09:54:35.1637626Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1637700Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1637744Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1637800Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1637942Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1638272Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1638348Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1638561Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1638650Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1638861Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1638952Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1639028Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1639073Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1639129Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1639228Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1639544Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1639617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1639828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1639933Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1640146Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1640248Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1640326Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1640367Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1640428Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1640530Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1640854Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1640929Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1641140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1641227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1641436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1641522Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1641596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1641637Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1641707Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1641816Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1642133Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1642207Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1642418Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1642506Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1642713Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1642804Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1642878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1642923Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1642982Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1643080Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1643399Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1643472Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1643681Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1643780Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1643997Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1644084Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1644157Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1644200Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1644255Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1644354Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1644680Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1644755Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1644965Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1645052Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1645259Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1645345Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1645417Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1645459Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1645525Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1645633Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1645950Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1646024Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1646235Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1646323Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1646531Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1646618Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1646694Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1646736Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1646849Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1646946Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1647267Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1647343Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1647553Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1647657Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1647882Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1647969Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1648042Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1648088Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1648143Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1648247Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1648566Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1648640Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1648849Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1648937Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1649145Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1649234Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1649305Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1649347Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1649426Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1649526Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1649844Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1649921Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1650129Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1650217Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1650427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1650515Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1650592Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1650634Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1650694Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1650794Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1651117Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1651191Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1651402Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1651499Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1651731Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1651816Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1651892Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1651935Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1651995Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1652093Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1652413Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1652490Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1652699Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1652788Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1652993Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1653082Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1653154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1653211Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1653281Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1653384Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1653703Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1653783Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1653993Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1654081Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1654288Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1654377Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1654451Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1654494Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1654549Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1654648Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1654966Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1655039Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1655249Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1655349Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1655572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1655658Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1655731Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1655772Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1655829Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1655928Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1656249Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1656324Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1656534Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1656622Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1656866Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1656955Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1657026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1657085Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1657157Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1657257Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1657671Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1657746Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1657957Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1658047Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1658255Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1658343Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1658415Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1658458Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1658515Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1658616Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1658931Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1659006Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1659214Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1659332Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1659544Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1659630Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1659703Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1659745Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1659802Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1659899Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1660225Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1660301Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1660510Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1660597Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1660804Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1660889Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1660972Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1661014Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1661082Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1661181Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1661497Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1661569Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1661777Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1661864Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1662070Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1662158Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1662233Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1662275Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1662332Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1662430Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1662744Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1662818Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1663026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1663135Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1663342Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1663429Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1663501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1663543Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1663599Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1663697Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1664016Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1664091Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1664300Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1664387Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1664595Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1664681Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1664764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1664816Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1664873Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1664971Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1665286Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1665359Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1665567Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1665653Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1665863Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1665950Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1666023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1666065Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1666121Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1666218Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1666538Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1666612Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1666861Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1666988Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1667194Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1667280Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1667353Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1667395Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1667451Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1667549Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1667865Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1667940Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1668148Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1668234Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1668441Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1668528Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1668614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1668670Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1668726Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1668825Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1669142Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1669216Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1669424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1669510Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1669719Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1669805Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1669878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1669919Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1669975Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1670073Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1670389Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1670462Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1670683Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1670779Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1670989Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1671075Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1671147Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1671189Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1671244Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1671344Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1671661Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1671735Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1671942Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1672029Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1672234Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1672321Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1672413Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1672457Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1672513Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1672611Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1672927Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1673000Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1673207Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1673296Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1673503Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1673591Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1673664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1673706Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1673761Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1673860Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1674178Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1674252Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1674472Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1674570Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1674776Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1674862Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1674935Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1674976Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1675032Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1675130Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1675447Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1675521Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1675728Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1675814Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1676026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1676123Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1676208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1676251Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1676308Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1676405Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1676721Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1676833Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1677041Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1677130Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1677338Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1677426Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1677497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1677540Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1677595Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1677693Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1678010Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1678087Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1678311Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1678412Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1678620Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1678707Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1678781Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1678822Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1678878Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1678977Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1679295Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1679369Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1679578Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1679665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1679872Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1679972Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1680063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1680106Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1680164Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1680261Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1680582Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1680655Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1680865Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1680953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1681159Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1681246Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1681327Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1681374Z Traceback (most recent call last): 2025-12-04T09:54:35.1681494Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1681533Z return value(self) 2025-12-04T09:54:35.1681681Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1681731Z self.assertExpectedInline( 2025-12-04T09:54:35.1681907Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1682061Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1682219Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1682263Z assert_expected_inline( 2025-12-04T09:54:35.1682409Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1682461Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1682627Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1682704Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1682821Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1682883Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1682977Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1683027Z raise self.failureException(msg) 2025-12-04T09:54:35.1683118Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1683240Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1683513Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1683649Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1683770Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1683869Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1684048Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1684050Z 2025-12-04T09:54:35.1684124Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1684275Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1684277Z 2025-12-04T09:54:35.1684365Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1684439Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1684482Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1684540Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1684641Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1684963Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1685037Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1685249Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1685336Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1685546Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1685648Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1685730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1685773Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1685829Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1685929Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1686245Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1686318Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1686527Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1686616Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1686867Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1686956Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1687028Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1687071Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1687126Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1687226Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1687573Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1687649Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1687856Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1687943Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1688150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1688236Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1688309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1688352Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1688411Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1688510Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1688826Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1688898Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1689109Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1689195Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1689403Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1689504Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1689590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1689631Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1689688Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1689786Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1690105Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1690179Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1690391Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1690480Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1690687Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1690774Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1690846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1690889Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1690945Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1691044Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1691390Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1691466Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1691675Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1691762Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1691969Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1692056Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1692128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1692171Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1692229Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1692328Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1692644Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1692718Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1692925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1693012Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1693221Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1697110Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1702069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1702110Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1702167Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1702264Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1702581Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1702653Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1702865Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1702952Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1703159Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1703245Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1703318Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1703359Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1703415Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1703513Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1703867Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1703947Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1704154Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1704242Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1704449Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1704537Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1704609Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1704653Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1704709Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1704811Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1705128Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1705200Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1705408Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1705495Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1705702Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1705818Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1705890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1705933Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1705988Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1706086Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1706404Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1706479Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1706689Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1706816Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1707022Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1707107Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1707180Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1707222Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1707278Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1707375Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1707718Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1707792Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1708001Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1708087Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1708295Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1708380Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1708454Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1708497Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1708554Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1708653Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1708971Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1709044Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1709253Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1709341Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1709548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1709664Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1709736Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1709778Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1709834Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1709932Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1710249Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1710324Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1710533Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1710622Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1710831Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1710919Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1710992Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1711035Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1711090Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1711189Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1711526Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1711601Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1711808Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1711894Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1712100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1712186Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1712262Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1712305Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1712361Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1712459Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1712777Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1712850Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1713060Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1713147Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1713354Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1713467Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1713539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1713581Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1713637Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1713734Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1714054Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1714127Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1714337Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1714424Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1714631Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1714718Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1714790Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1714833Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1714888Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1714997Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1715328Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1715403Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1715612Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1715699Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1715906Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1715993Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1716067Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1716109Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1716166Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1716263Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1716579Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1716652Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1716901Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1716988Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1717213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1717320Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1717394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1717435Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1717492Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1717589Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1717906Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1717981Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1718189Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1718279Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1718485Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1718571Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1718643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1718685Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1718741Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1718852Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1719184Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1719260Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1719468Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1719556Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1719763Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1719850Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1719923Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1719967Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1720024Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1720122Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1720441Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1720514Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1720723Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1720811Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1721033Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1721130Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1721203Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1721245Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1721300Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1721397Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1721719Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1721792Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1722004Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1722091Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1722298Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1722384Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1722460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1722501Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1722557Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1722675Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1722993Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1723067Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1723275Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1723362Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1723569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1723657Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1723730Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1723774Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1723830Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1723929Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1724246Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1724319Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1724525Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1724614Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1724835Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1724934Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1725006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1725049Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1725104Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1725202Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1725519Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1725594Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1725802Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1725888Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1726099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1726185Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1726258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1726299Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1726355Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1726473Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1726831Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1726904Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1727112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1727198Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1727404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1727491Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1727564Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1727607Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1727663Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1727761Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1728077Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1728151Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1728362Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1728450Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1728675Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1728775Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1728847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1728890Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1728946Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1729044Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1729360Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1729437Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1729645Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1729732Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1729937Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1730025Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1730096Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1730139Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1730195Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1730318Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1730641Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1730714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1730922Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1731008Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1731216Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1731302Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1731376Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1731419Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1731474Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1731572Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1731890Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1731963Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1732173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1732279Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1732487Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1732586Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1732659Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1732702Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1732757Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1732856Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1733173Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1733248Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1733456Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1733543Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1733749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1733836Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1733909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1733952Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1734019Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1734127Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1734445Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1734518Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1734725Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1734813Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1735024Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1735111Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1735185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1735228Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1735284Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1735380Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1735697Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1735768Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1735978Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1736078Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1736296Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1736382Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1736455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1736497Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1736553Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1736652Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1737004Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1737078Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1737288Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1737375Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1737582Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1737671Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1737752Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1737799Z Traceback (most recent call last): 2025-12-04T09:54:35.1737950Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1737992Z return value(self) 2025-12-04T09:54:35.1738142Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1738193Z self.assertExpectedInline( 2025-12-04T09:54:35.1738368Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1738505Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1738654Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1738699Z assert_expected_inline( 2025-12-04T09:54:35.1738845Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1738898Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1739064Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1739141Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1739259Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1739320Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1739413Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1741333Z raise self.failureException(msg) 2025-12-04T09:54:35.1741428Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1741548Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1741828Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1742005Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1742106Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1742204Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1742383Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1742385Z 2025-12-04T09:54:35.1742460Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1742612Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1742619Z 2025-12-04T09:54:35.1742708Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1742784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1742828Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1742885Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1742986Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1743308Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1743393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1743626Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1743718Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1743928Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1744013Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1744087Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1744129Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1744186Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1744284Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1744605Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1744684Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1744893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1744981Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1745191Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1745277Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1745350Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1745405Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1745461Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1745573Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1745891Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1745965Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1746172Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1746261Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1746467Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1746556Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1746629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1746670Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1746727Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1746869Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1747187Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1747273Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1747506Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1747596Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1747803Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1747890Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1747963Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1748005Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1748062Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1748161Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1748479Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1748553Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1748762Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1748848Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1749056Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1749142Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1749215Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1749275Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1749347Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1749448Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1749766Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1749839Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1750047Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1750135Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1750345Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1750432Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1750504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1750546Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1750601Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1750699Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1751014Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1751108Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1751319Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1751407Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1751617Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1751703Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1751776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1751818Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1751874Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1751973Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1752291Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1752364Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1752573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1752659Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1752867Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1752953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1753027Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1753080Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1753146Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1753244Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1753562Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1753636Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1753848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1753936Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1754147Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1754234Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1754306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1754348Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1754404Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1754502Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1754817Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1754910Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1755119Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1755208Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1755414Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1755501Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1755573Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1755615Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1755671Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1755770Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1756091Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1756165Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1756374Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1756460Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1756668Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1756795Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1756884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1756925Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1756994Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1757091Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1757409Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1757482Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1757691Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1757779Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1757987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1758075Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1758149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1758191Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1758248Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1758346Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1758682Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1758768Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1758977Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1759066Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1759272Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1759358Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1759430Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1759472Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1759528Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1759628Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1759944Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1760018Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1760225Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1760312Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1760520Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1760607Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1760691Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1760744Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1760800Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1760898Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1761214Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1761286Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1761495Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1761583Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1761790Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1761876Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1761950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1761991Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1762047Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1762144Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1762484Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1762569Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1762779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1762865Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1763075Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1763160Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1763232Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1763274Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1763330Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1763428Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1763747Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1763821Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1764029Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1764116Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1764322Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1764410Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1764495Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1764548Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1764604Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1764703Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1765020Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1765093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1765301Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1765390Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1765598Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1765686Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1765759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1765800Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1765856Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1765953Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1766292Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1766366Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1766576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1766665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1766923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1767009Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1767082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1767124Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1767182Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1767280Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1767601Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1767674Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1767882Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1767967Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1768176Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1768264Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1768363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1768418Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1768474Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1768572Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1768888Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1768963Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1769173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1769261Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1769467Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1769554Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1769626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1769668Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1769723Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1769821Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1770169Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1770244Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1770454Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1770542Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1770749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1770835Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1770907Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1770949Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1771006Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1771106Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1771430Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1771503Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1771712Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1771798Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1772006Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1772104Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1772178Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1772233Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1772289Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1772386Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1772703Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1772776Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1772987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1773076Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1773283Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1773370Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1773442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1773484Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1773540Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1773638Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1773977Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1774052Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1774260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1774347Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1774553Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1774639Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1774711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1774753Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1774809Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1774909Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1775230Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1775304Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1775513Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1775599Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1775808Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1775909Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1776004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1776046Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1776102Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1776199Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1776515Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1776587Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1776839Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1776927Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1777135Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1777221Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1777295Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1777336Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1777392Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1777489Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1777846Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1777924Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1778132Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1778219Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1778426Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1778513Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1778584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1778629Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1778685Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1778785Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1779101Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1779177Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1779385Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1779473Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1779683Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1779787Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1779872Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1779915Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1779971Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1780069Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1780386Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1780458Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1780668Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1780755Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1780963Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1781050Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1781123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1781164Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1781220Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1781317Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1781654Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1781729Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1781939Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1782025Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1782232Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1782318Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1782391Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1782433Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1782491Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1782589Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1782906Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1782982Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1783190Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1783277Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1783485Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1783588Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1783671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1783713Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1783769Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1783869Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1784187Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1784261Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1784469Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1784559Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1784773Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1784862Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1784935Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1784977Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1785033Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1785129Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1785468Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1785543Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1785751Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1785837Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1786047Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1786132Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1786206Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1786248Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1786306Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1786404Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1786721Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1786822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1787031Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1787117Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1787327Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1787435Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1787520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1787563Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1787618Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1787717Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1788034Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1788107Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1788318Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1788408Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1788614Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1788706Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1788786Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1788834Z Traceback (most recent call last): 2025-12-04T09:54:35.1788953Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1788993Z return value(self) 2025-12-04T09:54:35.1789156Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1789221Z self.assertExpectedInline( 2025-12-04T09:54:35.1789397Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1789537Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1789686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1789731Z assert_expected_inline( 2025-12-04T09:54:35.1789879Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1789932Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1790095Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1790173Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1790292Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1790355Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1790449Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1790497Z raise self.failureException(msg) 2025-12-04T09:54:35.1790589Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1790711Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1790989Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1791126Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1805415Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1805512Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1805691Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1805693Z 2025-12-04T09:54:35.1805767Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1805917Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1805919Z 2025-12-04T09:54:35.1806008Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1806087Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1806134Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1806190Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1806291Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1806612Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1806687Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1806912Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1807052Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1807261Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1807352Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1807424Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1807468Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1807522Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1807622Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1807942Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1808019Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1808230Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1808317Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1808522Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1808609Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1808682Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1808723Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1808780Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1808882Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1809228Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1809315Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1809523Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1809609Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1809817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1809905Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1809978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1810021Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1810077Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1810175Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1810493Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1810565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1810787Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1810884Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1811093Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1811181Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1811254Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1811296Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1811352Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1811451Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1811768Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1811843Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1812051Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1812141Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1812348Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1812434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1812507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1812549Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1812604Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1812705Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1813033Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1813120Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1813328Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1813416Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1813625Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1813712Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1813786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1813828Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1813884Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1813981Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1814297Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1814369Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1814587Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1814684Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1814893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1814980Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1815053Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1815095Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1815152Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1815249Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1815571Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1815644Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1815855Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1815943Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1816151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1816239Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1816311Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1816354Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1816410Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1816522Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1816888Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1816960Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1817167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1817255Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1817460Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1817548Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1817621Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1817665Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1817720Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1817819Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1818135Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1818208Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1818451Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1818539Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1818749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1818835Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1818908Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1818949Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1819006Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1819103Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1819423Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1819496Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1819708Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1819795Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1820004Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1820090Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1820163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1820205Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1820262Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1820377Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1820710Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1820785Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1820991Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1821079Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1821285Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1821374Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1821447Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1821489Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1821545Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1821644Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1821961Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1822036Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1822265Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1822354Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1822561Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1822648Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1822720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1822762Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1822818Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1822915Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1823234Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1823307Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1823517Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1823603Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1823810Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1823896Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1823969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1824011Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1824068Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1824178Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1824510Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1824583Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1824791Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1824877Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1825085Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1825175Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1825249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1825290Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1825346Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1825444Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1825759Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1825832Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1826061Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1826151Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1826359Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1826445Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1826518Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1826559Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1826615Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1826712Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1827057Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1827131Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1827338Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1827425Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1827630Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1827721Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1827794Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1827836Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1827894Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1828007Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1828339Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1828412Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1828623Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1828710Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1828925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1829012Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1829086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1829127Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1829183Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1829283Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1829604Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1829677Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1829912Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1830001Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1830208Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1830295Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1830366Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1830408Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1830463Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1830561Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1830881Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1830958Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1831166Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1831254Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1831460Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1831547Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1831619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1831662Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1831729Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1831827Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1832153Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1832227Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1832435Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1832522Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1832730Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1832817Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1832892Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1832933Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1832989Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1833087Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1833404Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1833476Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1833714Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1833802Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1834009Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1834095Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1834168Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1834209Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1834266Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1834364Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1834683Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1834757Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1834967Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1835054Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1835263Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1835349Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1835422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1835466Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1835534Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1835642Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1835957Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1836031Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1836239Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1836326Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1836533Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1836623Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1836695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1836736Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1836837Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1836936Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1837253Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1837343Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1837566Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1837655Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1837864Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1837949Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1838022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1838063Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1838120Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1838217Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1838536Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1838611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1838818Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1838904Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1839112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1839197Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1839271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1839335Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1839392Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1839503Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1839822Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1839895Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1840102Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1840188Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1840396Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1840493Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1840564Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1840608Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1840664Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1840762Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1841079Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1841172Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1841394Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1841485Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1841692Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1841780Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1841854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1841899Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1841955Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1842054Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1842375Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1842451Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1842663Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1842749Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1842958Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1843044Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1843121Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1843176Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1843234Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1843343Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1843661Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1843733Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1843944Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1844032Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1844241Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1844331Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1844408Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1844451Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1844511Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1844615Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1844936Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1845021Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1845241Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1845331Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1845537Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1845623Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1845697Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1845739Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1845795Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1845900Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1846225Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1846304Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1846512Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1846601Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1846844Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1846932Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1847009Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1847073Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1847154Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1847251Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1847571Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1847645Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1847854Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1847940Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1848152Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1848239Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1848315Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1848357Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1848417Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1848514Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1848834Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1848935Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1849147Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1849238Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1849446Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1849535Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1849608Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1849651Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1849707Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1849807Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1850126Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1850202Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1850411Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1850499Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1850709Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1850797Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1850871Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1850925Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1850994Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1851094Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1851413Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1851488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1851696Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1851785Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1851995Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1852084Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1852158Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1852199Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1852255Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1852353Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1852686Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1852771Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1852984Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1853071Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1853281Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1853369Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1853450Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1853497Z Traceback (most recent call last): 2025-12-04T09:54:35.1853619Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1853659Z return value(self) 2025-12-04T09:54:35.1853812Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1853861Z self.assertExpectedInline( 2025-12-04T09:54:35.1854036Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1854174Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1854324Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1854367Z assert_expected_inline( 2025-12-04T09:54:35.1854515Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1854569Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1854736Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1854825Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1854953Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1855014Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1855110Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1855160Z raise self.failureException(msg) 2025-12-04T09:54:35.1855250Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1855372Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1855648Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1855786Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1855887Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1855985Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1856163Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1856166Z 2025-12-04T09:54:35.1856241Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1856401Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1856406Z 2025-12-04T09:54:35.1856504Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1856580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1856622Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1856679Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1856824Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1857146Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1857219Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1857433Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1857521Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1857731Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1857817Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1857890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1857931Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1857988Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1858086Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1858405Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1858519Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1858728Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1858815Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1859023Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1859111Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1859183Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1859226Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1859283Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1859385Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1859704Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1859778Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1859987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1860076Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1860321Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1860409Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1860482Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1860525Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1860582Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1860681Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1860997Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1861071Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1861282Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1861373Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1861586Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1861672Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1861745Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1861786Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1861843Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1861941Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1862260Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1862364Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1862573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1862659Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1862866Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1862952Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1863026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1863068Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1863127Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1863225Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1863546Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1863620Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1863832Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1863920Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1864155Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1864244Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1864317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1864359Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1864414Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1864513Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1864830Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1864904Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1865114Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1865202Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1865408Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1865495Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1865567Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1865609Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1865664Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1865762Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1866086Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1866185Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1866393Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1866480Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1866686Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1866820Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1866894Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1866936Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1866995Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1867093Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1867413Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1867485Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1867694Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1867781Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1868018Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1868108Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1868182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1868225Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1868281Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1868380Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1868697Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1868770Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1868979Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1869070Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1869276Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1869363Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1869435Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1869476Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1869532Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1869629Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1869946Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1870044Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1870253Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1870341Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1870546Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1870633Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1870705Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1870751Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1870807Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1870906Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1871224Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1871296Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1871506Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1871592Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1871821Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1871909Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1871982Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1872023Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1872079Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1872177Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1872499Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1872572Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1872782Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1872871Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1873078Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1873163Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1873235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1873277Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1873334Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1873431Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1873749Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1873847Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1874055Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1874142Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1874348Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1874435Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1874509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1874553Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1874609Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1874708Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1875026Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1875099Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1875306Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1875404Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1875630Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1875719Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1875792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1875834Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1875890Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1875987Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1876304Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1876379Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1876589Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1876675Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1876922Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1877008Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1877080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1877121Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1877177Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1877275Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1877595Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1877700Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1877908Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1877994Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1878203Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1878290Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1878363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1878408Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1878464Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1878562Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1878880Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1878954Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1879164Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1879267Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1879486Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1879575Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1879647Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1879690Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1879745Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1879842Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1880158Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1880233Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1880442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1880532Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1880739Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1880825Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1880897Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1880938Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1880995Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1881096Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1885903Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1891440Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1891650Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1891736Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1891942Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1892030Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1892103Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1892145Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1892203Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1892301Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1892618Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1892693Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1892904Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1893019Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1893248Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1893337Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1893409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1893451Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1893506Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1893604Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1893919Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1893993Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1894202Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1894291Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1894495Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1894582Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1894653Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1894696Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1894751Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1894850Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1895185Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1895273Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1895484Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1895570Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1895779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1895866Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1895940Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1895981Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1896038Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1896136Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1896455Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1896529Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1896738Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1896877Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1897085Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1897172Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1897245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1897286Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1897342Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1897439Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1897762Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1897837Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1898045Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1898134Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1898340Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1898426Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1898498Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1898540Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1898596Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1898696Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1899035Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1899123Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1899331Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1899419Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1899625Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1899712Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1899787Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1899830Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1899885Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1899985Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1900304Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1900376Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1900603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1900700Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1900907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1900994Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1901066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1901107Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1901163Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1901261Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1901582Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1901657Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1901868Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1901954Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1902162Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1902248Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1902320Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1902363Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1902418Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1902518Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1902848Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1902933Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1903145Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1903233Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1903442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1903531Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1903604Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1903648Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1903703Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1903801Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1904120Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1904193Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1904412Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1904509Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1904716Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1904804Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1904877Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1904918Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1904975Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1905072Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1905396Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1905469Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1905678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1905764Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1905972Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1906057Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1906129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1906170Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1906227Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1906325Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1906651Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1906733Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1906976Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1907061Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1907269Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1907356Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1907429Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1907472Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1907527Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1907626Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1907942Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1908016Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1908251Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1908354Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1908561Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1908647Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1908719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1908761Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1908816Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1908914Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1909232Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1909305Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1909514Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1909600Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1909809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1909895Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1909968Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1910009Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1910066Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1910180Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1910509Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1910581Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1910790Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1910876Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1911084Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1911170Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1911244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1911286Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1911343Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1911440Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1911760Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1911832Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1912061Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1912150Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1912356Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1912443Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1912516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1912558Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1912613Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1912712Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1913030Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1913104Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1913312Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1913399Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1913605Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1913691Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1913763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1913806Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1913863Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1913976Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1914316Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1914390Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1914599Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1914686Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1914894Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1914983Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1915065Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1915112Z Traceback (most recent call last): 2025-12-04T09:54:35.1915234Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1915273Z return value(self) 2025-12-04T09:54:35.1915425Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1915475Z self.assertExpectedInline( 2025-12-04T09:54:35.1915650Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1915786Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1915960Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1916006Z assert_expected_inline( 2025-12-04T09:54:35.1916156Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1916209Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1916377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1916453Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1916573Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1916636Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1916731Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1916821Z raise self.failureException(msg) 2025-12-04T09:54:35.1916917Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1917039Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1917316Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1917453Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1917552Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1917653Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1917848Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1917880Z 2025-12-04T09:54:35.1917957Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1918105Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1918107Z 2025-12-04T09:54:35.1918198Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1918273Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1918318Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1918376Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1918477Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1918799Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1918876Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1919086Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1919175Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1919384Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1919470Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1919569Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1919612Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1919671Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1919774Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1920092Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1920164Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1920374Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1920465Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1920677Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1920768Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1920841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1920882Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1920941Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1921041Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1921360Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1921434Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1921656Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1921757Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1921963Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1922052Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1922128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1922171Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1922228Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1922329Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1922645Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1922722Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1922935Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1923026Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1923233Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1923325Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1923420Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1923467Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1923524Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1923622Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1923938Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1924013Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1924222Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1924311Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1924519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1924606Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1924678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1924719Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1924775Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1924876Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1925199Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1925272Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1925496Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1925597Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1925805Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1925891Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1925966Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1926007Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1926067Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1926167Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1926496Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1926570Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1926827Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1926915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1927122Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1927231Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1927318Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1927364Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1927420Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1927518Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1927835Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1927913Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1928123Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1928212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1928418Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1928508Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1928580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1928624Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1928679Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1928781Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1929100Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1929175Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1929413Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1929518Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1929728Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1929816Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1929892Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1929934Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1929990Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1930090Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1930412Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1932345Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1932560Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1932648Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1932855Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1932959Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1933050Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1933094Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1933154Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1933253Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1933587Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1933665Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1933875Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1933968Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1934177Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1934265Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1934340Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1934382Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1934440Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1934539Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1934859Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1934950Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1935162Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1935250Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1935459Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1935545Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1935623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1935665Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1935724Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1935828Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1936149Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1936283Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1936492Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1936583Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1936840Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1939302Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1939386Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1939429Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1939486Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1939583Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1939901Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1939974Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1940188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1940279Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1940489Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1940575Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1940649Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1940691Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1940748Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1940846Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1941165Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1941261Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1941472Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1941562Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1941768Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1941855Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1941926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1941969Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1942026Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1942128Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1942447Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1942547Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1942756Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1942842Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1943049Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1943157Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1943230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1943274Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1943330Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1943430Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1943751Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1943824Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1944033Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1944121Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1944328Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1944414Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1944488Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1944530Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1944586Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1944683Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1945001Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1945090Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1945299Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1945386Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1945593Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1945679Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1945755Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1945797Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1945855Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1945953Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1946271Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1946359Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1946567Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1946654Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1946924Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1947026Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1947100Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1947143Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1947199Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1947298Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1947614Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1947687Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1947899Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1947987Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1948196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1948284Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1948356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1948399Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1948454Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1948552Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1948871Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1948961Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1949171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1949258Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1949466Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1949552Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1949626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1949668Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1949725Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1949824Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1950142Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1950229Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1950439Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1950525Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1950747Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1950844Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1950919Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1950961Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1951018Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1951116Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1951435Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1951508Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1951718Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1951807Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1952016Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1952104Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1952176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1952219Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1952275Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1952376Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1952698Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1952786Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1952994Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1953081Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1953288Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1953374Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1953446Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1953489Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1953546Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1953646Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1953963Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1954050Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1954261Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1954348Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1954576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1954664Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1954738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1954780Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1954838Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1954936Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1955254Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1955328Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1955539Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1955628Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1955836Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1955924Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1955996Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1956039Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1956094Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1956192Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1956512Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1956602Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1956854Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1956942Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1957149Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1957236Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1957308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1957352Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1957409Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1957509Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1957825Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1957915Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1958122Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1958210Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1958442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1958530Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1958604Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1958645Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1958702Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1958799Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1959121Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1959194Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1959407Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1959494Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1959703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1959791Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1959864Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1959906Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1959962Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1960059Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1960380Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1960468Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1960679Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1960769Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1960978Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1961065Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1961138Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1961181Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1961238Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1961337Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1961655Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1961749Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1961957Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1962044Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1962273Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1962361Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1962434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1962476Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1962533Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1962631Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1962946Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1963021Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1963233Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1963321Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1963536Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1963622Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1963695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1963737Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1963793Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1963890Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1964209Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1964295Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1964505Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1964593Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1964801Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1964889Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1964962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1965005Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1965062Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1965160Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1965479Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1965567Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1965776Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1965863Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1966089Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1966177Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1966249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1966293Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1966348Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1966447Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1966797Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1966871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1967080Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1967167Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1967373Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1967461Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1967533Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1967575Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1967630Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1967728Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1968048Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1968138Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1968347Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1968434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1968641Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1968727Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1968802Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1968845Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1968901Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1968998Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1969316Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1969402Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1969612Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1969698Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1969932Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1970021Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1970094Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1970137Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1970199Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1970296Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1970617Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1970691Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1970902Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1970988Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1971194Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1971281Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1971353Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1971395Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1971450Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1971548Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1971866Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1971953Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1972163Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1972249Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1972455Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1972541Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1972622Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.1972672Z Traceback (most recent call last): 2025-12-04T09:54:35.1972797Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.1972837Z return value(self) 2025-12-04T09:54:35.1972989Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.1973053Z self.assertExpectedInline( 2025-12-04T09:54:35.1973231Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.1973368Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.1973518Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.1973575Z assert_expected_inline( 2025-12-04T09:54:35.1973738Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.1973793Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.1973958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.1974035Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.1974154Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.1974215Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.1974310Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.1974358Z raise self.failureException(msg) 2025-12-04T09:54:35.1974452Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.1974575Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.1974853Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.1974991Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.1975090Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.1975188Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.1975368Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.1975394Z 2025-12-04T09:54:35.1975471Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.1975619Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.1975621Z 2025-12-04T09:54:35.1975712Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.1975786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1975831Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1975888Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1975989Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1976309Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1976385Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1976598Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1976702Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1976946Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1977036Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1977108Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1977151Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1977221Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1977335Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1977654Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1977730Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1977940Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1978026Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1978235Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1978323Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1978397Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1978439Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1978496Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1978595Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1978913Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1978985Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1979194Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1979298Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1979507Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1979594Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1979669Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1979711Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1979767Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1979866Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1980185Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1980259Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1980467Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1980569Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1980775Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1980861Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1980933Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1980976Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1981043Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1981152Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1981470Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1981545Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1981755Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1981841Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1982048Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1982135Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1982208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1982249Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1982305Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1982403Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1982720Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1982792Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1983006Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1983103Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1983311Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1983398Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1983470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1983514Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1983568Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1983668Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1983985Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1984059Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1984269Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1984368Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1984574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1984660Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1984731Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1984774Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1984849Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1984949Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1985269Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1985343Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1985551Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1985638Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1985846Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1985933Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1986006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1986048Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1986104Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1986202Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1986518Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1986590Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1986842Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1986944Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1987151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1987238Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1987312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1987354Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1987410Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1987508Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1987828Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1987902Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1988110Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1988211Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1988416Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1988502Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1988574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1988629Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1988704Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1988804Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1989118Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1989192Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1989400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1989487Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1989696Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1989784Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1989856Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1989898Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1989954Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1990052Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1990368Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1990441Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1990651Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1990753Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1990959Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1991046Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1991117Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1991159Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1991215Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1991311Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1991637Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1991713Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1991922Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1992023Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1992230Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1992315Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1992387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1992442Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1992511Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1992609Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1992924Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1992998Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1993206Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1993293Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1993503Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1993590Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1993662Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1993704Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1993760Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1993858Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1994173Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1994245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1994454Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1994556Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1994762Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1994850Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1994921Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1994964Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1995020Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1995119Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1995438Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1995512Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1995723Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1995821Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1996027Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1996113Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1996197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1996239Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1996305Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1996402Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1996719Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1996839Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1997048Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1997133Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1997343Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1997430Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1997503Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1997544Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1997601Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1997700Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1998022Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1998095Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1998305Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1998412Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1998618Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1998706Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1998778Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.1998822Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.1998877Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.1998974Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.1999293Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.1999367Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.1999576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.1999685Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.1999896Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.1999986Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2000073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2000128Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2000184Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2000282Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2000598Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2000672Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2000880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2000967Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2001176Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2001264Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2001336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2001379Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2001434Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2001531Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2001853Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2001927Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2002135Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2002234Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2002442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2002530Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2002601Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2002643Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2002698Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2002797Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2003119Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2003194Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2003413Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2003501Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2003709Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2003798Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2003890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2003945Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2004000Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2004098Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2004413Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2004488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2004695Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2004783Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2004991Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2005078Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2005153Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2005195Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2005250Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2005347Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2005663Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2005736Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2005961Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2006047Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2006255Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2006342Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2006414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2006455Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2006511Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2006609Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2006961Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2007034Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2007258Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2007345Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2007551Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2007641Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2007738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2007782Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2007837Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2007935Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2008257Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2008331Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2008538Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2008625Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2008832Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2008920Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2008991Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2009038Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2009093Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2009191Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2009507Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2009582Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2009805Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2009891Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2010099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2010184Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2010259Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2010302Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2010358Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2010456Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2010773Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2010846Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2011064Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2011151Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2011358Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2011455Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2011539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2011582Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2011638Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2011735Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2012056Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2012130Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2012337Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2012427Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2012635Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2012722Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2012794Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2012837Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2012892Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2012991Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2013308Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2013382Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2013601Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2013688Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2013896Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2013982Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2014054Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2014098Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2014154Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2014253Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2014572Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2014646Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2014866Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2014952Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2015161Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2015258Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2015342Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2015386Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2015444Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2015541Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2015857Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2015930Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2016138Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2016225Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2016435Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2016521Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2016594Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2016635Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2016694Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2016835Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2017156Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2017230Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2017455Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2017543Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2017753Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2017839Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2017911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2017956Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2018010Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2018110Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2018427Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2018523Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2018733Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2018821Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2019026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2019130Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2019214Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2019257Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2019314Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2019414Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2019734Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2019806Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2020015Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2020102Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2020311Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2020396Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2020470Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2020511Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2020569Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2020667Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2020990Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2021075Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2021284Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2021370Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2021578Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2021664Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2021736Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2021780Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2021835Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2021937Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2022255Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2022341Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2022551Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2022639Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2022846Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2022955Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2023030Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2023074Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2023129Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2023228Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2023544Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2023617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2023827Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2023916Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2024126Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2024212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2024289Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2024330Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2024387Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2024484Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2024802Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2024890Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2025100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2025188Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2025397Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2025484Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2025560Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2025603Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2025661Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2025759Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2026080Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2026167Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2026376Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2026463Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2026669Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2026851Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2026936Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.2026985Z Traceback (most recent call last): 2025-12-04T09:54:35.2027108Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.2027148Z return value(self) 2025-12-04T09:54:35.2027299Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.2027354Z self.assertExpectedInline( 2025-12-04T09:54:35.2027530Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.2027670Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.2027821Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.2027867Z assert_expected_inline( 2025-12-04T09:54:35.2028015Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.2028067Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.2028232Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.2028310Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.2028429Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.2028494Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.2028587Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.2028641Z raise self.failureException(msg) 2025-12-04T09:54:35.2028735Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.2028872Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.2029152Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.2029289Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.2029389Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.2029486Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.2029671Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.2029674Z 2025-12-04T09:54:35.2029750Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2029898Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.2029917Z 2025-12-04T09:54:35.2030005Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2030081Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2030123Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2030184Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2030284Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2030630Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2030705Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2030918Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2031009Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2031219Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2031306Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2031380Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2031424Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2031481Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2031581Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2031899Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2031973Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2032181Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2032268Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2032475Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2032581Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2032652Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2032698Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2032754Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2032855Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2033173Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2033248Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2033457Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2033545Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2033753Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2033851Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2033924Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2033966Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2034022Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2034120Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2034460Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2034534Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2034747Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2034835Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2035045Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2035132Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2035208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2035250Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2035307Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2035404Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2035724Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2035798Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2036007Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2036095Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2036303Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2036404Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2036475Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2036518Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2036573Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2036671Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2037027Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2037102Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2037315Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2037403Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2037608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2037712Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2037784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2037826Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2037881Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2037979Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2038322Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2038397Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2038607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2038693Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2038901Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2038987Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2039062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2039104Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2039160Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2039260Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2039582Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2039655Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2039865Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2039953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2040163Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2040263Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2040336Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2040378Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2040435Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2040532Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2040847Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2040921Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2041131Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2041218Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2041435Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2041524Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2041597Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2041639Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2041694Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2041802Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2042129Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2042203Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2042412Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2042499Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2042703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2042791Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2042863Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2042907Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2042962Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2043059Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2043381Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2043453Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2043662Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2043749Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2043969Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2044054Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2044128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2044170Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2044226Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2044323Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2044639Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2044712Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2044921Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2045006Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2045224Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2045309Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2045382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2045423Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2045479Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2045587Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2045923Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2045996Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2046204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2046291Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2046496Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2046583Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2046655Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2046698Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2046790Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2046889Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2047206Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2047279Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2047486Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2047575Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2047804Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2047891Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2047965Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2048006Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2048063Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2048161Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2048478Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2048551Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2048761Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2048847Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2049068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2049154Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2049226Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2049268Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2049325Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2049452Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2049770Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2049843Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2050055Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2050140Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2050347Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2050434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2050506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2050549Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2050604Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2050701Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2051018Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2051090Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2051299Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2051387Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2051604Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2051691Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2051764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2051808Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2051863Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2051961Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2052279Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2052353Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2052562Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2052649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2052867Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2052953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2053027Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2053068Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2053124Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2053241Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2053560Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2053633Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2053841Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2053927Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2054139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2054226Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2054300Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2054341Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2054398Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2054495Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2054815Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2054887Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2055098Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2055187Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2055403Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2055490Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2055563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2055606Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2055661Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2055760Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2056078Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2056154Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2056364Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2056465Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2056671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2056799Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2056872Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2056914Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2056969Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2057095Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2057413Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2057487Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2057696Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2057782Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2057988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2058075Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2058148Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2058189Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2058244Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2058342Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2058665Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2058737Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2058946Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2059046Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2059252Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2059338Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2059412Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2059453Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2059508Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2059606Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2059926Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2060000Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2060210Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2060314Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2060520Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2060607Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2060679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2060722Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2060796Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2060905Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2061225Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2061299Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2061505Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2061591Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2061796Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2061883Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2061957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2061998Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2062054Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2062153Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2062470Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2062543Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2062752Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2062850Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2063060Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2063147Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2063222Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2063263Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2063319Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2063416Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2063737Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2063811Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2064019Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2064116Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2064323Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2064409Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2064482Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2064523Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2064590Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2064701Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2065020Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2065095Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2065303Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2065390Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2065596Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2065684Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2065756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2065798Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2065853Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2065952Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2066268Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2066342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2066552Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2066651Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2066913Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2067001Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2067075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2067116Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2067173Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2067270Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2067593Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2067666Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2067875Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2067980Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2068186Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2068271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2068344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2068384Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2068467Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2068567Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2068888Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2068964Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2069172Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2069258Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2069468Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2069557Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2069629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2069671Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2069727Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2069826Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2070142Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2070215Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2070424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2070525Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2070730Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2070818Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2070891Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2070934Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2070990Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2071089Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2071409Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2071483Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2071691Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2071789Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2071996Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2072082Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2072154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2072206Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2072274Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2072373Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2072690Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2072762Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2072969Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2073056Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2073266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2073354Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2073426Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2073467Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2073525Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2073621Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2073937Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2074009Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2074218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2074318Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2074524Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2074612Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2074684Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2074726Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2074781Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2074879Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2075197Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2075272Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2075484Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2075591Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2075798Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2075884Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2075956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2076009Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2076076Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2076175Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2076491Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2076565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2076812Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2076899Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2077107Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2077194Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2077268Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2077309Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2077367Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2077468Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2077789Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2077862Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2078074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2078174Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2078380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2078467Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2078540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2078581Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2078637Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2078734Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2079053Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2079126Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2079332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2079434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2079642Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2079729Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2079813Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2079868Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2079926Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2080025Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2080342Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2080416Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2080625Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2080712Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2080921Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2081009Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2081081Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2081123Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2081180Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2081278Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2081597Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2081671Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2081886Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2081986Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2082193Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2082280Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2082361Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.2082407Z Traceback (most recent call last): 2025-12-04T09:54:35.2082530Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.2082568Z return value(self) 2025-12-04T09:54:35.2082721Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.2082770Z self.assertExpectedInline( 2025-12-04T09:54:35.2082946Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.2083082Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.2083241Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.2083283Z assert_expected_inline( 2025-12-04T09:54:35.2083431Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.2083482Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.2083659Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.2083744Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.2083864Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.2083924Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.2084018Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.2084067Z raise self.failureException(msg) 2025-12-04T09:54:35.2084160Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.2084280Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.2084561Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.2084703Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.2084802Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.2084900Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.2085076Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.2085078Z 2025-12-04T09:54:35.2085154Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2085302Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.2085304Z 2025-12-04T09:54:35.2085393Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2085483Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2085528Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2085584Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2085684Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2086003Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2086078Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2086292Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2086380Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2086588Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2086674Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2086800Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2086842Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2086898Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2086997Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2087330Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2087417Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2087626Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2087712Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2087921Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2088007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2088080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2088121Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2088178Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2088277Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2088599Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2088674Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2088881Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2088968Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2089172Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2089261Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2089352Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2089394Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2089450Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2089549Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2089867Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2089941Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2090149Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2090239Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2090444Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2090531Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2090618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2090660Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2090715Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2090813Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2091151Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2091225Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2091434Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2091522Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2091730Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2091816Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2091890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2091931Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2091990Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2092088Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2092408Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2092482Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2092692Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2092779Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2092988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2093074Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2093159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2093202Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2093257Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2093357Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2093675Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2093748Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2093956Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2094046Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2094252Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2094339Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2094422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2094464Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2094519Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2094618Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2094961Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2095036Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2095243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2095332Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2095538Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2095626Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2095699Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2095740Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2095798Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2095897Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2096214Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2096287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2096495Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2096581Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2096831Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2096932Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2097006Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2097048Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2097105Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2097203Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2097521Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2097593Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2097803Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2097890Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2098097Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2098198Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2098269Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2098312Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2098367Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2098465Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2098804Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2098879Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2099086Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2099174Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2099382Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2099470Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2099541Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2099584Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2099641Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2099740Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2100057Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2100130Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2100338Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2100425Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2100633Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2100732Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2100806Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2100847Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2100903Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2101001Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2101320Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2101391Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2101602Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2101689Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2101897Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2101996Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2102069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2102109Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2102166Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2102263Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2102603Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2102677Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2102886Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2102974Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2103180Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2103267Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2103338Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2103381Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2103438Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2103537Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2103857Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2103931Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2104138Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2104225Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2104431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2104532Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2104605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2104647Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2104704Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2104802Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2105120Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2105194Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2105405Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2105492Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2105700Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2105799Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2105872Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2105913Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2105970Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2106067Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2106408Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2106482Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2106690Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2106827Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2107034Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2107120Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2107193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2107234Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2107293Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2107391Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2107709Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2107783Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2107994Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2108081Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2108287Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2108393Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2108465Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2108675Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2110736Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2110838Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2111157Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2111230Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2111441Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2111532Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2111744Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2111859Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2111931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2111974Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2112030Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2112130Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2112484Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2112560Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2112767Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2112856Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2113061Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2113149Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2113222Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2113265Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2113322Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2113421Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2113740Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2113815Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2114023Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2114110Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2114320Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2114420Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2114493Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2114535Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2114593Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2114691Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2115009Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2115081Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2115291Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2115378Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2115585Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2115683Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2115756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2115798Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2115854Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2115952Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2116289Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2116363Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2116570Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2116658Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2116900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2116986Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2117061Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2117105Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2117161Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2117259Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2117577Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2117651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2117858Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2117945Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2118153Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2118271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2118344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2118387Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2118444Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2118542Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2118862Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2118936Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2119147Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2119233Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2119441Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2119543Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2119616Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2119657Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2119713Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2119811Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2120152Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2120226Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2120436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2120523Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2120731Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2120818Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2120890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2120935Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2120990Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2121088Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2121408Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2121483Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2121691Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2121778Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2121986Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2122088Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2122161Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2122205Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2122260Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2122359Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2122676Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2122751Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2122960Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2123047Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2123255Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2123356Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2123428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2123470Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2123527Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2123625Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2123971Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2124046Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2124255Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2124342Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2124550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2124636Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2124711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2124754Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2124810Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2124907Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2125227Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2125300Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2125512Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2125600Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2125809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2125910Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2125982Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2126026Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2126081Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2126181Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2126499Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2126574Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2126812Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2126899Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2127120Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2127206Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2127278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2127321Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2127376Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2127490Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2127822Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2127898Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2128108Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2128194Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2128402Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2128488Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2128563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2128605Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2128661Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2128758Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2129078Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2129150Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2129359Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2129446Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2129671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2129758Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2129833Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2129876Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2129932Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2130030Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2130347Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2130422Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2130631Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2130719Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2130939Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2131026Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2131097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2131140Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2131196Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2131308Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2131641Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2131715Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2131923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2132010Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2132217Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2132304Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2132376Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2132420Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2132475Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2132573Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2132892Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2132965Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2133173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2133260Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2133479Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2133565Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2133641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2133683Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2133739Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2133836Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2134160Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2134234Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2134444Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2134530Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2134750Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2134836Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2134909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2134950Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2135006Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2135124Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2135443Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2135519Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2135727Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2135814Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2136021Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2136109Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2136183Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2136226Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2136282Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2136384Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2136708Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2136819Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2137026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2137114Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2137345Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2137432Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2137506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2137547Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2137603Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2137700Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2138020Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2138094Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2138302Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2138388Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2138612Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2138697Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2138770Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2138812Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2138867Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2138996Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2139316Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2139390Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2139600Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2139686Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2139898Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2139987Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2140060Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2140103Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2140158Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2140257Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2140576Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2140651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2140862Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2140950Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2143803Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2143889Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2143970Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.2144019Z Traceback (most recent call last): 2025-12-04T09:54:35.2144142Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.2144182Z return value(self) 2025-12-04T09:54:35.2144332Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.2144383Z self.assertExpectedInline( 2025-12-04T09:54:35.2144561Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.2144700Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.2144848Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.2153596Z assert_expected_inline( 2025-12-04T09:54:35.2153742Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.2153795Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.2153958Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.2154035Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.2154180Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.2154265Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.2154360Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.2154408Z raise self.failureException(msg) 2025-12-04T09:54:35.2154501Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.2154623Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.2154902Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.2155039Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.2155140Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.2155237Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.2155416Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.2155420Z 2025-12-04T09:54:35.2155496Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2155645Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.2155647Z 2025-12-04T09:54:35.2155736Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2155811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2155855Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2155929Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2156030Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2156354Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2156430Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2156642Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2156731Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2156974Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2157063Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2157136Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2157179Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2157250Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2157349Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2157664Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2157737Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2157977Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2158067Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2158273Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2158361Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2158434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2158475Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2158532Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2158630Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2158950Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2159023Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2159230Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2159317Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2159524Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2159610Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2159683Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2159726Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2159801Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2159898Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2160217Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2160290Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2160497Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2160583Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2160793Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2160882Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2160954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2161008Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2161064Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2161162Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2161478Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2161564Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2161784Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2161873Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2162079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2162167Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2162239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2162281Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2162337Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2162435Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2162753Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2162827Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2163037Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2163125Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2163332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2163418Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2163491Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2163547Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2163603Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2163699Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2164017Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2164090Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2164299Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2164386Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2164594Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2164682Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2164754Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2164809Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2164863Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2164961Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2165282Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2165366Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2165588Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2165677Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2165882Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2165970Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2166042Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2166084Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2166139Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2166238Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2166556Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2166630Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2166878Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2166964Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2167171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2167256Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2167333Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2167391Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2167447Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2167545Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2167862Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2167935Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2168142Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2168229Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2168437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2168523Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2168596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2168652Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2168708Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2168805Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2169120Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2169208Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2169431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2169519Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2169727Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2169815Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2169887Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2169929Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2169984Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2170082Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2170400Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2170474Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2170681Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2170769Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2170974Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2171061Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2171134Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2171187Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2171242Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2171340Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2171661Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2171733Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2171941Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2172028Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2172236Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2172322Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2172395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2172454Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2172510Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2172607Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2172922Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2173013Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2173224Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2173311Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2173519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2173605Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2173677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2173719Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2173776Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2173874Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2174196Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2174269Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2174477Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2174564Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2174770Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2174858Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2174932Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2174986Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2175041Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2175138Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2175454Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2175527Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2175734Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2175822Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2176030Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2176118Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2176203Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2176245Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2176300Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2176398Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2176732Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2176856Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2177067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2177153Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2177361Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2177448Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2177520Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2177562Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2177618Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2177719Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2178039Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2178112Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2178324Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2178410Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2178618Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2178706Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2178794Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2178836Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2178892Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2178991Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2179310Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2179383Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2179593Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2179681Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2179889Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2179975Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2180061Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2180103Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2180158Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2180256Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2180597Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2180681Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2180890Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2180976Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2181184Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2181270Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2181348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2181390Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2181446Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2181545Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2181863Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2181936Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2182144Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2182230Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2182436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2182523Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2182609Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2182651Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2182707Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2182805Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2183125Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2183197Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2183409Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2183497Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2183704Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2183791Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2183881Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2183923Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2183978Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2184076Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2184405Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2184492Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2184700Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2184787Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2184995Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2185081Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2185152Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2185194Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2185249Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2185349Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2185666Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2185741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2185947Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2186037Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2186243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2186330Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2186422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2186464Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2186520Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2186616Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2186970Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2187042Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2187251Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2187339Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2187548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2187633Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2187722Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2187763Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2187820Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2187918Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2188262Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2188337Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2188545Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2188633Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2188840Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2188927Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2189000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2189043Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2189100Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2189202Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2189521Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2189979Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2190308Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2190648Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2190983Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2191319Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2191532Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2191685Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2191814Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2192007Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2192469Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2192902Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2193226Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2193560Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2193895Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2194225Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2194434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2194588Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2194718Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2194909Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2195383Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2195813Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2196135Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2196466Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2196835Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2197168Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2197363Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2197514Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2197641Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2197833Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2198292Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2198724Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2199145Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2199487Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2199821Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2200175Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2200372Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2200527Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2200654Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2200846Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2201303Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2201730Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2202053Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2202386Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2202718Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2203078Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2203274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2203425Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2203555Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2203747Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2204232Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2204661Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2204984Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2205318Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2205653Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2205987Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2206184Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2206336Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2206463Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2206652Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2207151Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2207578Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2207899Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2208233Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2208567Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2208915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2209111Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2209262Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2209391Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2209581Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2210036Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2210467Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2210793Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2211124Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2211459Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2211806Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2212004Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2212155Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2212281Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2212474Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2212964Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2213393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2213715Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2214051Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2214391Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2214722Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2214917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2215068Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2215195Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2215386Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2215840Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2216268Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2216589Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2216957Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2217290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2217635Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2217829Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2217981Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2218108Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2218297Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2218756Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2219180Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2219507Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2219837Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2220168Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2220512Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2220706Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2220855Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2220981Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2221170Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2221651Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2222078Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2222396Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2222728Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2223058Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2223388Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2223584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2223736Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2223864Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2224054Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2224508Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2224938Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2225257Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2225586Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2225918Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2226258Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2226453Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2226603Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2226733Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2226967Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2227421Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2227846Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2228168Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2228498Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2228831Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2229177Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2229373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2229523Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2229649Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2229837Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2230316Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2230743Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2231062Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2231394Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2231725Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2232057Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2232254Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2232405Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2232532Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2232720Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2233175Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2233604Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2233925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2234255Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2234587Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2234934Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2235129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2235295Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2235421Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2235609Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2236064Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2236490Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2236857Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2237187Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2237518Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2237860Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2238065Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.2238228Z Traceback (most recent call last): 2025-12-04T09:54:35.2238427Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.2238619Z return value(self) 2025-12-04T09:54:35.2238849Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.2239098Z self.assertExpectedInline( 2025-12-04T09:54:35.2239353Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.2239699Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.2240017Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.2240242Z assert_expected_inline( 2025-12-04T09:54:35.2240458Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.2240690Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.2240939Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.2241211Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.2241438Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.2241653Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.2241837Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.2242014Z raise self.failureException(msg) 2025-12-04T09:54:35.2242187Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.2242436Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.2242868Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.2243317Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.2243605Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.2243839Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.2244150Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.2244364Z 2025-12-04T09:54:35.2244439Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2244697Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.2244881Z 2025-12-04T09:54:35.2244974Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2245173Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2245326Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2245455Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2245648Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2246114Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2246543Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2246933Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2247281Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2247616Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2247946Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2248144Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2248294Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2248422Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2248612Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2249071Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2249501Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2249822Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2252703Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2253038Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2253366Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2253562Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2253713Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2253839Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2254031Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2254508Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2254936Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2255255Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2255584Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2255916Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2256248Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2256443Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2256593Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2256718Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2256943Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2257413Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2257836Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2258170Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2258514Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2258848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2259178Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2259375Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2259527Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2259652Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2259840Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2260295Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2260723Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2261043Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2261374Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2261703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2262032Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2262226Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2262376Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2262502Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2262690Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2263158Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2263584Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2263903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2264231Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2264562Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2264892Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2265087Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2265237Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2265363Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2265551Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2266028Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2266454Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2266851Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2267198Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2267532Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2267863Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2268060Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2268212Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2268338Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2268527Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2268983Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2269408Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2269728Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2270058Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2270387Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2270715Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2270909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2271059Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2271186Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2271396Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2271849Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2272278Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2272598Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2272927Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2273258Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2273589Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2273786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2273936Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2274062Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2274267Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2274721Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2275146Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2275486Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2275818Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2276151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2276481Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2276674Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2276873Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2277000Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2277189Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2277646Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2278071Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2278389Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2278721Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2279052Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2279382Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2279576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2279731Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2279859Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2280067Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2280525Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2280953Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2281273Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2281604Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2281937Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2282269Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2282466Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2282619Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2282749Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2282961Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2283417Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2283849Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2284199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2284531Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2284864Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2285193Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2285389Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2285539Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2285664Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2285857Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2286316Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2286775Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2287095Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2287427Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2287757Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2288091Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2288287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2288441Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2288571Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2288775Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2289233Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2289660Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2289984Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2290317Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2290655Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2290993Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2291192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2291346Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2291473Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2291678Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2292133Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2292562Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2292914Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2293245Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2293577Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2293909Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2294105Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2294257Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2294386Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2294574Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2295031Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2295465Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2295794Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2296129Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2296461Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2296827Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2297026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2297176Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2297320Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2297510Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2297965Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2298393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2298713Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2299047Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2299379Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2299709Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2299903Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2300057Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2300203Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2300394Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2300851Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2301279Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2301629Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2301962Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2302297Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2302631Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2302828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2302980Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2303107Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2303300Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2306395Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2306865Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2307191Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2307526Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2307862Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2308191Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2308388Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2308542Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2311206Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2311397Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2311853Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2312281Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2312601Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2312929Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2313260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2313591Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2313787Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2313938Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2322440Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2322628Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2323081Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2323506Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2323860Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2324192Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2324523Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2324853Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2325048Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2325198Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2325324Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2325511Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2325967Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2326396Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2326716Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2327064Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2327394Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2327722Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2327918Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2328068Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2328218Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2328407Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2328864Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2329289Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2329607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2329937Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2330267Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2330597Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2330793Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2330957Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2331082Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2331270Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2331723Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2332163Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2332500Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2332833Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2333164Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2333495Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2333691Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2333843Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2333968Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2334157Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2334611Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2335041Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2335362Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2335691Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2336021Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2336348Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2336543Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2336709Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2336871Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2337059Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2337513Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2337942Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2338261Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2338591Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2338922Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2339254Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2339450Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2339617Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2339742Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2339930Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2340383Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2340822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2341158Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2341488Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2341819Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2342150Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2342348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2342499Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2342625Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2342815Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2343272Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2343699Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2344020Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2344350Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2344679Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2345007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2345204Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2345369Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2345495Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2345683Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2346136Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2346562Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2346914Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2347244Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2347575Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2347904Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2348098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2348265Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2348390Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2348577Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2349029Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2349465Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2349799Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2350130Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2350339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2350425Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2350498Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2350540Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2350597Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2350696Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2351018Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2351092Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2351302Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2351389Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2351597Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2351685Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2351759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2351815Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2351870Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2351969Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2352287Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2352360Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2352567Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2352655Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2352862Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2352948Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2353021Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2353082Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2353138Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2353237Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2353553Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2353655Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2353864Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2353952Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2354160Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2354245Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2354317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2354358Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2354416Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2354513Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2354834Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2354906Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2355116Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2355202Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2355411Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2355498Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2355572Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2355625Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2355681Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2355778Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2356097Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2356169Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2356377Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2356464Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2356671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2356796Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2356882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2356924Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2356981Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2357079Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2357407Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2357495Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2357709Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2357797Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2358004Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2358090Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2358162Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2358204Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2358260Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2358360Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2358676Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2358749Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2358959Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2359045Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2359252Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2359339Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2359426Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2359467Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2359524Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2359621Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2359940Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2360012Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2360221Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2360308Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2360515Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2360601Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2360686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2360728Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2360784Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2360881Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2361213Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2361299Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2361508Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2361595Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2361802Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2361890Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2361962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2362004Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2362059Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2362158Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2362474Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2362548Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2362753Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2362841Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2363047Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2363134Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2363216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2363258Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2363316Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2363413Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2363732Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2363804Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2364014Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2364102Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2364310Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2364397Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2364478Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2364520Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2364576Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2364674Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2365001Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2365085Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2365294Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2365383Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2365591Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2365678Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2365751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2365793Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2365849Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2365948Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2366269Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2366343Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2366553Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2366639Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2366877Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2366963Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2367060Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.2367109Z Traceback (most recent call last): 2025-12-04T09:54:35.2367234Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.2367274Z return value(self) 2025-12-04T09:54:35.2367428Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.2367478Z self.assertExpectedInline( 2025-12-04T09:54:35.2367655Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.2367792Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.2367943Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.2367987Z assert_expected_inline( 2025-12-04T09:54:35.2368137Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.2368188Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.2368369Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.2368445Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.2368564Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.2368625Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.2368719Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.2368768Z raise self.failureException(msg) 2025-12-04T09:54:35.2368894Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.2369017Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.2369295Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.2369435Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.2369535Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.2369633Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.2369815Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.2369819Z 2025-12-04T09:54:35.2369898Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2370048Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.2370051Z 2025-12-04T09:54:35.2370141Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2370216Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2370260Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2370316Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2370417Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2370739Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2370825Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2371039Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2371137Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2371345Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2371432Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2371506Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2371549Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2371606Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2371706Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2372023Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2372106Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2372318Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2372404Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2372633Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2372721Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2372795Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2372838Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2372894Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2372992Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2373308Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2373382Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2373593Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2373682Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2373889Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2373977Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2374049Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2374091Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2374147Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2374244Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2374563Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2374646Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2374853Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2374941Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2375147Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2375233Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2375306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2375350Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2375406Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2375504Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2375819Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2375903Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2376112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2376198Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2376424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2376511Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2376584Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2376626Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2376682Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2376808Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2377125Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2377198Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2377408Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2377493Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2377699Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2377785Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2377857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2377898Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2377954Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2378052Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2378372Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2378462Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2378670Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2378763Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2378968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2379055Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2379127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2379170Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2379226Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2379324Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2379642Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2379732Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2379940Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2380039Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2380258Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2380347Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2380419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2380462Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2380518Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2380616Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2380936Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2381009Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2381217Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2381303Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2381509Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2381595Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2381668Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2381709Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2381765Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2381863Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2382180Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2382263Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2382475Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2382560Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2382767Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2382852Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2382926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2382968Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2383024Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2383121Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2383438Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2383527Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2383733Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2383829Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2384047Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2384134Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2384206Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2384250Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2384305Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2384404Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2384723Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2384799Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2385009Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2385096Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2385302Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2385389Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2385463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2385503Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2385559Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2385657Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2385986Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2386058Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2386267Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2386353Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2386559Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2386646Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2386721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2386795Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2386853Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2386949Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2387283Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2387355Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2387565Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2387663Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2387884Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2387972Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2388044Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2388087Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2388143Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2388241Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2388556Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2388631Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2388844Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2388932Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2389144Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2389232Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2389305Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2389347Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2389403Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2389503Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2389832Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2389905Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2390114Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2390201Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2390407Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2390494Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2390568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2390609Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2390665Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2390762Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2391092Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2391165Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2391375Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2391480Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2391688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2391774Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2391848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2391890Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2391949Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2392046Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2392367Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2392440Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2392651Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2392737Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2392944Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2393031Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2393104Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2393146Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2393201Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2393303Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2393634Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2393708Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2393917Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2394004Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2394210Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2394299Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2394373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2394414Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2394469Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2394569Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2394895Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2394968Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2395187Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2395286Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2395494Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2395579Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2395653Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2395694Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2395754Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2395852Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2396169Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2396242Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2396450Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2396536Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2396780Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2396867Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2396939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2396980Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2397038Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2397136Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2397477Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2397552Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2397760Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2397850Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2398057Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2398145Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2398217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2398261Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2398317Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2398419Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2398751Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2398824Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2399043Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2399151Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2399358Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2399444Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2399517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2399559Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2399614Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2399715Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2400039Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2400113Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2400322Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2400410Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2400618Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2400703Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2400775Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2400816Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2400877Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2400974Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2401304Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2401379Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2401588Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2401674Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2401881Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2401969Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2402044Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2402086Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2402142Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2402250Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2402568Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2402643Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2402859Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2402961Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2403167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2403254Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2403327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2403369Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2403425Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2403525Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2403843Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2403918Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2404125Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2404216Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2404424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2404513Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2404590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2404632Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2404691Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2404799Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2405116Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2405188Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2405398Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2405485Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2405695Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2405782Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2405855Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2405898Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2405958Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2406066Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2406386Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2406460Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2406691Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2406806Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2407018Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2407108Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2407182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2407225Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2407280Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2407380Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2407699Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2407775Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2407983Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2408075Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2408281Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2408368Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2408439Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2408481Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2408537Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2408652Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2408974Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2409051Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2409258Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2409348Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2409558Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2409649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2409724Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2409767Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2409827Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2409941Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2410259Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2410332Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2410572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2410662Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2410876Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2410964Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2411038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2411079Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2411137Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2411236Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2411558Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2411633Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2411843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2411935Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2412140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2412226Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2412298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2412342Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2412399Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2412518Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2412832Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2412909Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2413118Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2413208Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2413418Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2413508Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2413579Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2413623Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2413691Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2413790Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2414107Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2414180Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2414407Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2414498Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2414708Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2414795Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2414868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2414909Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2414965Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2415062Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2415383Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2415457Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2415666Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2415753Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2415963Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2416050Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2416125Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2416167Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2416234Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2416331Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2416649Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2416725Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2416976Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2417067Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2417274Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2417365Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2417438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2417482Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2417552Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2417650Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2417969Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2418043Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2418277Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2418367Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2418572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2418660Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2418732Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2418774Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2418829Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2418928Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2419246Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2419320Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2419528Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2419615Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2419824Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2419910Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2419984Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2420025Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2420095Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2420193Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2420510Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2420584Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2420791Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2420876Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2421086Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2421172Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2421244Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2421285Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2421353Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2421450Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2421769Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2421843Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2422069Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2422158Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2422364Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2422452Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2422525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2422566Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2422622Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2422720Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2423040Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2423114Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2423324Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2423413Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2423622Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2423710Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2423784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2423830Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2423897Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2423995Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2424311Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2424383Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2424591Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2424676Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2424885Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2424971Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2425043Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2425095Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2425151Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2425248Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2425564Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2425653Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2425874Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2425961Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2426170Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2426259Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2426331Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2426373Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2426430Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2426529Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2426876Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2426950Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2427159Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2427245Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2427451Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2427537Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2427619Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.2427684Z Traceback (most recent call last): 2025-12-04T09:54:35.2427805Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.2427845Z return value(self) 2025-12-04T09:54:35.2427996Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.2428047Z self.assertExpectedInline( 2025-12-04T09:54:35.2428222Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.2428360Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.2428510Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.2428553Z assert_expected_inline( 2025-12-04T09:54:35.2428702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.2428754Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.2428919Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.2429008Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.2429127Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.2429187Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.2429280Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.2429328Z raise self.failureException(msg) 2025-12-04T09:54:35.2429421Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.2429567Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.2429844Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.2429980Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.2430081Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.2430177Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.2430357Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.2430359Z 2025-12-04T09:54:35.2430436Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2430585Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.2430588Z 2025-12-04T09:54:35.2430675Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2430752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2430794Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2430851Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2430951Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2431273Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2431360Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2431571Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2431662Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2431870Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2431957Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2432029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2432072Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2432129Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2432229Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2432548Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2432632Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2432843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2432931Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2433137Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2433244Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2433319Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2433360Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2433417Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2433516Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2433838Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2433912Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2434122Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2434211Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2434419Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2434506Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2434580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2434621Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2434678Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2434776Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2435096Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2435179Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2435387Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2435475Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2435682Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2435770Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2435842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2435884Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2435941Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2436042Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2436359Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2436443Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2436651Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2436738Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2437007Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2437108Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2437183Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2437225Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2437280Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2437379Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2437703Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2437777Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2437987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2438075Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2438283Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2438369Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2438441Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2438483Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2438539Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2438637Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2438956Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2439042Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2439251Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2439338Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2439546Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2439633Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2439707Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2439748Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2439805Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2439904Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2440226Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2440318Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2440527Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2440614Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2440829Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2440928Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2441001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2441042Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2441097Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2441196Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2441512Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2441586Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2441793Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2441880Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2442085Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2442173Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2442245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2442287Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2442342Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2442440Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2442759Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2442843Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2443052Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2443139Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2443346Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2443431Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2443504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2443546Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2443603Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2443701Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2444019Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2444102Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2444312Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2444397Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2444623Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2444711Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2444783Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2444824Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2444880Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2444978Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2445296Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2445369Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2445578Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2445667Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2445872Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2445959Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2446031Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2446072Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2446128Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2446225Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2446542Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2446628Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2446873Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2446962Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2447168Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2447255Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2447328Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2447370Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2447428Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2447527Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2447848Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2447941Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2448148Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2448234Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2448466Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2448553Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2448626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2448666Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2448724Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2448822Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2449143Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2449215Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2449427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2449514Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2449720Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2449807Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2449879Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2449921Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2449976Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2450075Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2450393Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2450495Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2450703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2450791Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2450996Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2451084Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2451156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2451198Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2451255Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2451352Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2451668Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2451761Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2451969Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2452057Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2452285Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2452372Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2452446Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2452487Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2452544Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2452642Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2452960Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2453032Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2453246Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2453334Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2453542Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2453629Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2453701Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2453742Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2453799Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2453897Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2454217Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2454306Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2454515Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2454604Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2454809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2454896Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2454969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2455012Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2455069Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2455167Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2455486Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2455573Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2455782Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2455869Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2456094Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2456183Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2456255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2456298Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2456354Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2456452Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2456808Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2456883Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2457093Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2457181Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2457388Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2457474Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2457550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2457591Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2457647Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2457745Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2458067Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2458159Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2458368Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2458456Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2458662Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2458748Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2458823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2458865Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2458922Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2459019Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2459337Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2459426Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2459634Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2459721Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2459951Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2460040Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2460112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2460154Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2460210Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2460307Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2460623Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2460697Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2460906Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2460993Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2461199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2461287Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2461359Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2461402Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2461458Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2461556Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2461873Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2461956Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2462166Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2462252Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2462457Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2462543Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2462617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2462659Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2462715Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2462811Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2463130Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2463213Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2463421Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2463519Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2463736Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2463823Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2463897Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2463939Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2463997Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2464095Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2464416Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2464491Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2464702Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2464790Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2464998Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2465084Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2465156Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2465199Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2465254Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2465352Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2465669Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2465756Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2465966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2466054Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2466262Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2466350Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2466425Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2466467Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2466522Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2466620Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2466990Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2467063Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2467271Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2467370Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2467598Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2467683Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2467757Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2467799Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2467855Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2467952Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2468270Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2468345Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2468557Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2468642Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2468852Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2468938Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2469010Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2469052Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2469107Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2469205Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2469534Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2469608Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2469817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2469903Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2470110Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2470197Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2470270Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2470312Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2470367Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2470465Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2470799Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2470872Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2471079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2471174Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2471394Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2471480Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2471554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2471595Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2471651Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2471748Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2472067Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2472141Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2472350Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2472435Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2472642Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2472728Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2472800Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2472841Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2472898Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2472997Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2473325Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2473398Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2473607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2473694Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2473899Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2473987Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2474060Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2474102Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2474158Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2474255Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2474585Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2474658Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2474866Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2474977Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2475185Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2475272Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2475345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2475387Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2475443Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2475540Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2475857Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2475932Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2476141Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2476227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2476434Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2476519Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2476592Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2476633Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2476689Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2476826Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2477159Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2477233Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2477442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2477527Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2477735Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2477821Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2477895Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2477937Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2477994Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2478091Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2478429Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2478503Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2478724Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2478827Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2479035Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2479122Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2479196Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2479238Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2479294Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2479392Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2479710Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2479786Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2479995Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2480082Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2480290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2480376Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2480448Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2480490Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2480545Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2480645Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2480971Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2481047Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2481255Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2481341Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2481550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2481637Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2481711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2481752Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2481808Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2481905Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2482232Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2482305Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2482531Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2482628Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2482840Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2482926Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2483000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2483041Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2483096Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2483193Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2483513Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2483587Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2483795Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2483884Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2484089Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2484175Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2484247Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2484289Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2484345Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2484445Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2484771Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2484846Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2485056Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2485144Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2485352Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2485439Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2485513Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2485555Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2485611Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2485721Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2486037Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2486110Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2486330Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2486429Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2486637Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2486724Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2486838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2486879Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2486935Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2487032Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2487350Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2487423Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2487633Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2487720Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2487928Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2488014Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2488086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2488128Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2488184Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2488297Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2488613Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2488687Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2488893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2488980Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2489188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2489276Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2489357Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.2489405Z Traceback (most recent call last): 2025-12-04T09:54:35.2489526Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.2489578Z return value(self) 2025-12-04T09:54:35.2489728Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.2489778Z self.assertExpectedInline( 2025-12-04T09:54:35.2489953Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.2490091Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.2490266Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.2490344Z assert_expected_inline( 2025-12-04T09:54:35.2490502Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.2490585Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.2490765Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.2490866Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.2493486Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.2493554Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.2493650Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.2493701Z raise self.failureException(msg) 2025-12-04T09:54:35.2493798Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.2493922Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.2494202Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.2494341Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.2494440Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.2494538Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.2494719Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.2494739Z 2025-12-04T09:54:35.2494820Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2494970Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.2494975Z 2025-12-04T09:54:35.2495063Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2495141Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2495183Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2495240Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2495339Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2495663Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2495737Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2495961Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2496049Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2496258Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2496343Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2496428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2496493Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2496551Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2496649Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2497002Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2497078Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2497289Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2497377Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2497586Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2497673Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2497746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2497790Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2497845Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2497944Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2498262Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2498336Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2498562Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2498649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2498855Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2498943Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2499015Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2499058Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2499113Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2499214Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2499539Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2499612Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2499838Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2499923Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2500131Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2500217Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2500321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2500364Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2500420Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2500517Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2500834Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2500906Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2501115Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2501202Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2501410Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2501498Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2501571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2501613Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2501669Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2501766Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2502083Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2502158Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2502380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2502469Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2502676Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2502763Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2502835Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2502877Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2502932Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2503031Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2503349Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2503424Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2503644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2503732Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2503938Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2504036Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2504118Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2504161Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2504216Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2504315Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2504635Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2504709Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2504921Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2505008Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2505218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2505303Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2505377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2505420Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2505477Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2505575Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2505898Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2505973Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2506193Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2506280Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2506488Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2506573Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2506646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2506687Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2506778Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2506876Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2507194Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2507267Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2507493Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2507580Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2507785Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2507886Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2507976Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2508021Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2508076Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2508174Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2508492Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2508565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2508773Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2508860Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2509067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2509153Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2509225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2509269Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2509325Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2509422Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2509739Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2509813Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2510034Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2510119Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2510327Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2510413Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2510486Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2510527Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2510583Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2510684Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2511002Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2511084Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2511293Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2511379Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2511586Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2511682Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2511766Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2511810Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2511865Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2511963Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2512281Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2512353Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2512562Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2512650Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2512858Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2512944Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2513017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2513058Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2513114Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2513211Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2513528Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2513613Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2513820Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2513906Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2514113Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2514198Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2514271Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2514311Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2514367Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2514465Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2514784Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2514877Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2515088Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2515174Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2515380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2515486Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2515561Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2515602Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2515659Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2515755Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2516074Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2516146Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2516355Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2516443Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2516649Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2516737Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2516849Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2516891Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2516947Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2517045Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2517366Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2517454Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2517661Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2517749Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2517954Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2518040Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2518112Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2518153Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2518211Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2518310Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2518627Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2518713Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2518921Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2519007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2519225Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2519324Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2519399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2519440Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2519497Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2519595Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2519912Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2519984Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2520192Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2520280Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2520486Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2520571Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2520645Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2520686Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2520742Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2520839Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2521158Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2521242Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2521449Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2521537Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2521745Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2521832Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2521903Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2521945Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2522002Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2522101Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2522416Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2522506Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2522714Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2522801Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2523020Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2523117Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2523191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2523232Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2523288Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2523388Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2523707Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2523783Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2523994Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2524081Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2524288Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2524375Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2524448Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2524488Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2524545Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2524642Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2524963Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2525047Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2525255Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2525342Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2525548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2525633Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2525705Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2525747Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2525805Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2525904Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2526223Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2526306Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2526514Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2526600Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2526872Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2526972Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2527044Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2527086Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2527142Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2527241Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2527558Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2527632Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2527841Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2527929Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2528137Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2528225Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2528297Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2528339Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2528394Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2528492Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2528809Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2528901Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2529109Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2529197Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2529404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2529489Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2529561Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2529603Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2529660Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2529758Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2530075Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2530161Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2530371Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2530456Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2530682Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2530769Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2530841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2530883Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2530938Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2531037Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2531355Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2531428Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2531638Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2531726Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2531932Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2532019Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2532091Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2532133Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2532188Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2532286Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2532604Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2532689Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2532897Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2532985Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2533191Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2533278Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2533351Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2533393Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2533451Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2533547Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2533865Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2533947Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2534160Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2534247Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2534476Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2534563Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2534637Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2534677Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2534735Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2534833Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2535154Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2535227Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2535436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2535524Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2535730Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2535818Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2535890Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2535932Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2535988Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2536086Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2536407Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2536494Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2536702Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2536822Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2537028Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2537115Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2537186Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2537229Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2537287Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2537386Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2537703Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2537795Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2538003Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2538089Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2538319Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2538406Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2538479Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2538520Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2538578Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2538675Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2538992Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2539066Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2539275Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2539362Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2539569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2539656Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2539728Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2539770Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2539825Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2539923Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2540242Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2540330Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2540537Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2540625Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2540832Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2540919Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2540992Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2541036Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2541093Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2541191Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2541507Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2541595Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2541803Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2541890Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2542124Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2542213Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2542287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2542329Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2542386Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2542482Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2542800Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2542874Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2543087Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2543174Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2543381Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2543468Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2543540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2543581Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2543638Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2543735Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2544054Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2544143Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2544355Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2544442Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2544649Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2544736Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2544809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2544853Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2544909Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2545009Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2545329Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2545414Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2545621Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2545707Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2545937Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2546024Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2546096Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2546139Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2546195Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2546292Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2546609Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2546683Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2546937Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2547023Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2547230Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2547316Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2547388Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2547429Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2547485Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2547582Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2547902Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2551155Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2551367Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2551453Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2551660Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2551745Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2551819Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2551861Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2551917Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2552013Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2552330Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2561014Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2561222Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2561329Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2561548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2561636Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2561708Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2561751Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2561807Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2561904Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2562220Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2562296Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2562505Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2562592Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2562799Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2562885Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2562957Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2562999Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2563054Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2563155Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2563474Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2563566Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2563776Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2563862Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2564068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2564153Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2564228Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2564270Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2564326Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2564424Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2564755Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2564827Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2565035Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2565133Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2565354Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2565440Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2565514Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2565556Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2565612Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2565709Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2566026Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2566101Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2566309Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2566401Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2566607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2566693Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2566812Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.2566860Z Traceback (most recent call last): 2025-12-04T09:54:35.2566982Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.2567024Z return value(self) 2025-12-04T09:54:35.2567176Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.2567245Z self.assertExpectedInline( 2025-12-04T09:54:35.2567421Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.2567561Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.2567708Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.2567751Z assert_expected_inline( 2025-12-04T09:54:35.2567898Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.2567950Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.2568117Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.2568194Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.2568311Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.2568373Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.2568480Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.2568529Z raise self.failureException(msg) 2025-12-04T09:54:35.2568620Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.2568745Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.2569054Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.2569192Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.2569293Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.2569391Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.2569570Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.2569573Z 2025-12-04T09:54:35.2569647Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2569798Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.2569800Z 2025-12-04T09:54:35.2569889Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2569964Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2570006Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2570064Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2570164Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2570492Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2570567Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2570781Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2570883Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2571090Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2571179Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2571251Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2571293Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2571349Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2571448Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2571768Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2571842Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2572050Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2572151Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2572356Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2572443Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2572515Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2572570Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2572636Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2572735Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2573051Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2573125Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2573332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2573418Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2573626Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2573713Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2573786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2573827Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2573883Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2573981Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2574298Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2574370Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2574579Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2574682Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2574888Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2574975Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2575047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2575088Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2575145Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2575242Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2575565Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2575640Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2575847Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2575946Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2576151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2576237Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2576308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2576366Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2576435Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2576535Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2576886Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2576960Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2577166Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2577254Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2577462Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2577549Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2577623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2577667Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2577724Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2577822Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2578137Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2578210Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2578420Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2578527Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2578735Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2578823Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2578896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2578938Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2578993Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2579090Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2579410Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2579483Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2579691Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2579790Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2579998Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2580085Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2580157Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2580211Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2580281Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2580380Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2580694Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2580768Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2580974Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2581061Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2581267Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2581355Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2581426Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2581468Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2581525Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2581626Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2581945Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2582019Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2582231Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2582332Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2582539Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2582626Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2582699Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2582740Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2582796Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2582894Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2583213Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2583287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2583495Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2583600Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2583806Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2583892Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2583976Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2584028Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2584087Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2584184Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2584503Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2584577Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2584785Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2584877Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2585083Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2585171Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2585243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2585286Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2585341Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2585440Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2585758Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2585832Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2586041Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2586141Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2586347Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2586434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2586507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2586549Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2586604Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2586703Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2587061Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2587135Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2587342Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2587445Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2587652Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2587738Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2587823Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2587884Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2587941Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2588039Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2588356Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2588430Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2588640Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2588726Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2588936Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2589022Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2589100Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2589144Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2589201Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2589299Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2589620Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2589696Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2589919Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2590008Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2590217Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2590309Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2590384Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2590428Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2590483Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2590581Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2590902Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2590979Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2591199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2591289Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2591495Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2591581Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2591664Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2591720Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2591778Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2591877Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2592192Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2592268Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2592478Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2592566Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2592775Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2592862Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2592935Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2592979Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2593038Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2593135Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2593452Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2593525Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2593747Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2593835Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2594044Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2594132Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2594209Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2594251Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2594312Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2594410Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2594730Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2594805Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2595025Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2595117Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2595324Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2595414Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2595507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2595552Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2595609Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2595709Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2596027Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2596101Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2596307Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2596396Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2596603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2596690Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2596807Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2596851Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2596907Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2597006Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2597325Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2597399Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2597630Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2597717Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2597925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2598011Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2598086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2598127Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2598184Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2598282Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2598604Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2598677Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2598900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2598986Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2599192Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2599295Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2599379Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2599426Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2599481Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2599581Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2599899Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2599973Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2600182Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2600270Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2600476Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2600563Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2600636Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2600680Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2600736Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2600834Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2601149Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2601225Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2601442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2601529Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2601737Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2601823Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2601899Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2601940Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2601997Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2602095Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2602416Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2602488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2602709Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2602794Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2603004Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2603101Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2603185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2603227Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2603283Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2603381Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2603698Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2603771Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2603980Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2604069Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2604276Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2604363Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2604436Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2604478Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2604534Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2604632Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2604948Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2605022Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2605240Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2605328Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2605534Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2605621Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2605693Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2605736Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2605792Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2605891Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2606211Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2606302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2606508Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2606594Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2606834Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2606935Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2607022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2607063Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2607119Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2607218Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2607535Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2607606Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2607814Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2607900Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2608108Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2608193Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2608267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2608308Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2608364Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2608460Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2608777Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2608865Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2609073Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2609162Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2609369Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2609456Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2609528Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2609571Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2609626Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2609727Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2610046Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2610135Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2610343Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2610433Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2610640Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2610754Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2610828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2610870Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2610925Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2611024Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2611343Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2611417Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2611627Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2611716Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2611925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2612010Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2612084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2612125Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2612181Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2612278Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2612600Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2612684Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2612893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2612979Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2613188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2613274Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2613347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2613388Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2613445Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2613543Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2613860Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2613943Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2614151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2614238Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2614454Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2614550Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2614624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2614666Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2614721Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2614820Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2615137Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2615211Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2615419Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2615508Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2615714Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2615800Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2615873Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2615916Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2615971Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2616069Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2616386Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2616470Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2616678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2616889Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2617099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2617185Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2617258Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2617299Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2617356Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2617455Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2617771Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2617863Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2618071Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2618157Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2618375Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2618475Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2618547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2618591Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2618646Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2618745Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2619060Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2619134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2619345Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2619433Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2619638Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2619725Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2619796Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2619838Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2619894Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2619991Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2620309Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2620397Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2620602Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2620690Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2620896Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2620982Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2621055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2621096Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2621153Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2621251Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2621570Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2621653Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2621861Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2621947Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2622163Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2622259Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2622332Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2622374Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2622434Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2622542Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2622858Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2622930Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2623139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2623226Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2623431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2623519Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2623590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2623634Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2623690Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2623787Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2624104Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2624186Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2624393Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2624481Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2624687Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2624774Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2624845Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2624888Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2624943Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2625042Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2625359Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2625442Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2625650Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2625736Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2625967Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2626055Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2626127Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2626168Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2626225Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2626322Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2626639Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2626712Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2626957Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2627043Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2627249Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2627336Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2627409Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2627450Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2627507Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2627603Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2627923Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2628015Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2628222Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2628311Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2628517Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2628604Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2628676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2628720Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2628777Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2628875Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2629191Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2629277Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2629487Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2629573Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2629803Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2629891Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2629964Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2630006Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2630062Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2630160Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2630476Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2630550Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2630761Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2630848Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2631054Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2631141Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2631223Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.2631269Z Traceback (most recent call last): 2025-12-04T09:54:35.2631390Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.2631428Z return value(self) 2025-12-04T09:54:35.2631583Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.2631645Z self.assertExpectedInline( 2025-12-04T09:54:35.2631825Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.2631961Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.2632111Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.2632153Z assert_expected_inline( 2025-12-04T09:54:35.2632301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.2632352Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.2632516Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.2632594Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.2632714Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.2632775Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.2632883Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.2632940Z raise self.failureException(msg) 2025-12-04T09:54:35.2633034Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.2633154Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.2633430Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.2633586Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.2633686Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.2633784Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.2633962Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.2633964Z 2025-12-04T09:54:35.2634047Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2634195Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.2634198Z 2025-12-04T09:54:35.2634288Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2634362Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2634406Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2634462Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2634564Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2634883Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2634957Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2635170Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2635264Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2635488Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2635576Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2635652Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2635694Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2635752Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2635849Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2636168Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2636244Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2636452Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2636539Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2636793Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2636880Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2636954Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2636995Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2637051Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2637176Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2637495Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2637569Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2637777Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2637864Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2638070Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2638159Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2638233Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2638276Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2638332Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2638431Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2638746Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2638819Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2639026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2639116Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2639338Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2639427Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2639499Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2639542Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2639598Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2639697Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2640013Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2640088Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2640296Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2640384Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2640613Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2640700Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2640772Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2640813Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2640869Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2640988Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2641308Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2641383Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2641591Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2641677Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2641885Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2641973Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2642047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2642088Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2642144Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2642241Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2642560Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2642634Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2642843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2642931Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2643150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2643236Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2643309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2643351Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2643406Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2643504Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2643822Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2643896Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2644103Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2644201Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2644408Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2644496Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2644568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2644610Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2644674Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2644784Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2645102Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2645175Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2645385Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2645471Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2645678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2645765Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2645839Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2645880Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2645936Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2646033Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2646351Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2646424Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2646634Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2646732Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2646983Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2647070Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2647143Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2647185Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2647240Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2647339Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2647658Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2647732Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2647939Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2648041Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2648246Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2648332Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2648404Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2648446Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2648513Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2648626Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2648942Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2649017Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2649226Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2649312Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2649519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2649606Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2649679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2649720Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2649776Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2649873Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2650189Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2650260Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2650470Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2650570Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2650777Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2650864Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2650937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2650978Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2651035Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2651131Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2651453Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2651528Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2651739Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2651837Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2652042Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2652129Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2652201Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2652242Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2652309Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2652424Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2652743Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2652821Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2653028Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2653117Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2653325Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2653416Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2653489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2653532Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2653590Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2653690Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2654006Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2654080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2654293Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2654402Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2654608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2654696Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2654771Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2654814Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2654872Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2654969Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2655289Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2655364Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2655575Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2655673Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2655885Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2655972Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2656047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2656102Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2656171Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2656271Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2656588Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2656664Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2656911Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2657000Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2657208Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2657300Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2657371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2657413Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2657469Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2657567Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2657884Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2657957Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2658167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2658272Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2658479Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2658567Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2658640Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2658684Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2658741Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2658842Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2659160Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2659236Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2659447Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2659548Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2659757Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2659842Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2659917Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2659973Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2660045Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2660145Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2660461Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2660536Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2660746Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2660834Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2661044Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2661131Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2661205Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2661248Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2661308Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2661404Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2661719Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2661795Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2662005Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2662105Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2662312Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2662401Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2662473Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2662518Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2662574Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2662678Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2662995Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2663073Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2663279Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2663379Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2663584Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2663671Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2663743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2663794Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2663860Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2663960Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2664281Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2664354Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2664563Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2664649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2664856Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2664943Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2665017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2665059Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2665119Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2665218Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2665537Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2665609Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2665818Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2665914Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2666125Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2666216Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2666290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2666333Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2666388Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2666488Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2666843Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2666921Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2667128Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2667232Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2667441Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2667529Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2667614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2667680Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2667739Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2667838Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2668156Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2668233Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2668442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2668530Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2668741Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2668828Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2668902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2668946Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2669002Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2669098Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2669416Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2669489Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2669701Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2669802Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2670009Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2670096Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2670170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2670213Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2670272Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2670370Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2670695Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2670770Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2670981Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2671083Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2671291Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2671380Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2671462Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2671519Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2671575Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2671675Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2671994Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2672071Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2672280Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2672369Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2672576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2672665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2672739Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2672787Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2672844Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2672946Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2673264Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2673342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2673565Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2673652Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2673860Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2673948Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2674024Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2674066Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2674122Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2674219Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2674537Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2674611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2677230Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2677327Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2677537Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2677624Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2677744Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2677788Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2677845Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2677943Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2678261Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2678334Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2678544Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2678630Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2678839Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2678926Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2678999Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2679042Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2679099Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2679196Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2679512Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2679594Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2679824Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2679912Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2680118Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2680205Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2680276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2680318Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2680375Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2680474Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2680792Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2680866Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2681086Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2681173Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2681379Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2681466Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2681561Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2681605Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2681662Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2681759Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2682079Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2682151Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2682359Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2682446Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2682654Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2682740Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2682813Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2682855Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2682910Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2683007Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2683325Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2683398Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2683624Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2683711Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2683920Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2684007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2684079Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2684121Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2684177Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2684275Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2684593Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2684665Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2684884Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2684971Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2685178Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2685273Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2685355Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2685398Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2685454Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2685551Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2685867Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2685940Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2686149Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2686237Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2686445Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2686531Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2686605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2686646Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2686703Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2686830Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2687150Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2687223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2687449Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2687534Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2687742Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2687827Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2687901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2687941Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2687997Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2688096Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2688416Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2688488Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2688710Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2688797Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2689003Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2689103Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2689191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2689233Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2689288Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2689386Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2689703Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2689777Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2689984Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2690072Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2690278Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2690365Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2690438Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2690480Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2690535Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2690637Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2690955Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2691042Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2691250Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2691336Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2691544Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2691629Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2691702Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2691743Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2691799Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2691897Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2692215Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2692298Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2692506Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2692592Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2692801Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2692896Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2692980Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2693022Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2693078Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2693175Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2693492Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2693565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2693774Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2693861Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2694067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2694154Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2694227Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2694269Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2694324Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2694423Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2694740Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2694826Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2695035Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2695122Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2695329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2695415Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2695487Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2695529Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2695584Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2695685Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2696003Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2696086Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2696294Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2696380Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2696586Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2696699Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2696826Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2696868Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2696924Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2697022Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2697344Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2697415Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2697623Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2697711Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2697919Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2698005Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2698080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2698121Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2698178Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2698275Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2698592Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2698681Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2698889Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2698977Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2699184Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2699271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2699353Z _________________ GPUTests.test_remove_noop_slice_scatter_cuda _________________ 2025-12-04T09:54:35.2699401Z Traceback (most recent call last): 2025-12-04T09:54:35.2699526Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 14842, in new_test 2025-12-04T09:54:35.2699568Z return value(self) 2025-12-04T09:54:35.2699721Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 6762, in test_remove_noop_slice_scatter 2025-12-04T09:54:35.2699772Z self.assertExpectedInline( 2025-12-04T09:54:35.2699951Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 3272, in assertExpectedInline 2025-12-04T09:54:35.2700104Z return super().assertExpectedInline(actual if isinstance(actual, str) else str(actual), expect, skip + 1) 2025-12-04T09:54:35.2700254Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 413, in assertExpectedInline 2025-12-04T09:54:35.2700298Z assert_expected_inline( 2025-12-04T09:54:35.2700446Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 378, in assert_expected_inline 2025-12-04T09:54:35.2700510Z assert_eq(expect, actual, msg=help_text) 2025-12-04T09:54:35.2700688Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/expecttest/__init__.py", line 450, in assertMultiLineEqualMaybeCppStack 2025-12-04T09:54:35.2700766Z self.assertMultiLineEqual(expect, actual, *args, **kwargs) 2025-12-04T09:54:35.2700883Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1226, in assertMultiLineEqual 2025-12-04T09:54:35.2700946Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T09:54:35.2701040Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T09:54:35.2701089Z raise self.failureException(msg) 2025-12-04T09:54:35.2701181Z AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.2701304Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.2701585Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.2701722Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.2701822Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.2701925Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.2702108Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.2702111Z 2025-12-04T09:54:35.2702191Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2702344Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.2702357Z 2025-12-04T09:54:35.2702446Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2702521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2702564Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2702623Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2702722Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2703043Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2703119Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2703331Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2703425Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2703637Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2703741Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2703815Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2703857Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2703913Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2704012Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2704352Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2704427Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2704636Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2704725Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2704932Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2705019Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2705091Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2705136Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2705192Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2705292Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2705610Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2705684Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2705893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2705980Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2706186Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2706287Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2706361Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2706403Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2706459Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2706556Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2706921Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2706995Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2707204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2707291Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2707498Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2707606Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2707680Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2707721Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2707777Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2707874Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2708221Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2708296Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2708504Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2708591Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2708797Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2708884Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2708958Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2709003Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2709059Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2709158Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2709478Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2709555Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2709762Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2709851Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2710058Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2710162Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2710235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2710277Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2710334Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2710432Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2710751Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2710825Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2711037Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2711124Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2711334Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2721055Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2721129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2721171Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2721227Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2729690Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2730031Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2730105Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2730316Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2730402Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2730610Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2730697Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2730771Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2730813Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2730869Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2730967Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2731286Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2731362Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2731572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2731660Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2731884Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2731971Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2732043Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2732086Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2732142Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2732241Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2732558Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2732633Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2732841Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2732929Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2733153Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2733239Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2733314Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2733356Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2733414Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2733529Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2733873Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2733946Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2734158Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2734245Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2734454Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2734543Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2734618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2734660Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2734718Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2734814Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2735134Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2735207Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2735417Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2735506Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2735731Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2735821Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2735894Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2735937Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2735992Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2736090Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2736407Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2736482Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2736690Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2736804Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2737028Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2737115Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2737189Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2737230Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2737286Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2737398Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2737732Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2737806Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2738020Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2738106Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2738315Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2738401Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2738476Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2738517Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2738574Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2738670Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2738990Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2739063Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2739272Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2739360Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2739585Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2739672Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2739745Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2739786Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2739841Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2739939Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2740258Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2740333Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2740543Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2740630Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2740848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2740936Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2741008Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2741050Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2741106Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2741227Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2741545Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2741620Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2741827Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2741915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2742120Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2742206Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2742282Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2742323Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2742379Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2742476Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2742796Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2742868Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2743078Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2743165Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2743388Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2743474Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2743548Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2743589Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2743645Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2743742Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2744060Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2744134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2744343Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2744429Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2744658Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2744746Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2744817Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2744859Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2744914Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2745032Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2745351Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2745425Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2745634Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2745721Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2745926Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2746014Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2746087Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2746129Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2746185Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2746283Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2746601Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2746676Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2746929Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2747021Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2747246Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2747333Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2747407Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2747448Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2747506Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2747602Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2747926Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2747999Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2748209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2748310Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2748519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2748605Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2748678Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2748719Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2748789Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2748900Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2749218Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2749294Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2749501Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2749587Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2749792Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2749880Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2749953Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2749995Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2750051Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2750150Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2750466Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2750539Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2750748Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2750848Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2751053Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2751140Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2751214Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2751255Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2751312Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2751410Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2751730Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2751803Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2752010Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2752109Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2752316Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2752402Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2752475Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2752516Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2752587Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2752695Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2753016Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2753090Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2753299Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2753386Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2753595Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2753683Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2753758Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2753799Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2753855Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2753954Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2754272Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2754346Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2754555Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2754655Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2754861Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2754950Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2755023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2755066Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2755121Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2755221Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2755540Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2755616Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2755823Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2755923Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2756130Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2756217Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2756290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2756331Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2756397Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2756505Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2756866Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2756940Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2757149Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2757234Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2757443Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2757531Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2757603Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2757644Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2757699Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2757797Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2758116Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2758188Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2758396Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2758504Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2758712Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2758800Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2758872Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2758913Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2758969Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2759066Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2759387Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2759462Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2759671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2759771Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2759977Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2760063Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2760136Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2760192Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2760262Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2760363Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2760677Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2760752Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2760960Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2761048Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2761255Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2761342Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2761414Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2761456Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2761517Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2761617Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2761934Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2762006Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2762215Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2762313Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2762522Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2762610Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2762684Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2762725Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2762781Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2762880Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2763199Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2763272Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2763482Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2763582Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2763788Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2763875Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2763947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2763999Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2764067Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2764166Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2764482Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2764557Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2764763Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2764850Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2765058Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2765146Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2765218Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2765260Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2765316Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2765416Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2765733Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2765810Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2766023Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2766124Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2766332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2766419Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2766492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2766533Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2766591Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2766688Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2767049Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2767123Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2767333Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2767442Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2767652Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2767739Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2767828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2767870Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2767944Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2768041Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2768359Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2768436Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2768644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2768732Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2768939Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2769027Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2769098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2769140Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2769197Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2769295Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2769611Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2769684Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2769893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2769997Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2770202Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2770290Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2770364Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2770407Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2770462Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2770561Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2770881Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2770955Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2771164Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2771263Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2771471Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2771556Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2771641Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2771703Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2771761Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2771858Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2772176Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2772251Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2772460Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2772546Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2772758Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2772845Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2772920Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2772963Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2773019Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2773115Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2773437Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2773513Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2773720Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2773825Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2774032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2774121Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2774195Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2774237Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2774294Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2774394Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2774713Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2774791Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2775011Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2775100Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2775305Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2775397Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2775480Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2775535Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2775595Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2775692Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2776012Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2776085Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2776294Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2776382Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2776593Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2776681Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2776800Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2776842Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2776899Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2776998Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2777319Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2777394Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2777621Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2777710Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2777915Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2778003Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2778075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2778116Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2778171Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2778269Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2778590Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2778666Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2778885Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2778973Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2779180Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2779272Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2779372Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2779420Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2779478Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2779578Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2779894Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2779973Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2780180Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2780268Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2780477Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2780564Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2780640Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2780684Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2780743Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2780841Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2781159Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2781234Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2781457Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2781544Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2781753Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2781840Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2781915Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2781956Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2782015Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2782113Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2782431Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2782504Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2782725Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2782813Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2783018Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2783104Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2783195Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2783240Z frames [('total', 1), ('ok', 1)] 2025-12-04T09:54:35.2783297Z stats [('calls_captured', 4), ('unique_graphs', 1)] 2025-12-04T09:54:35.2783394Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2783710Z inductor [('triton_bundler_save_kernel', 24), ('benchmarking.InductorBenchmarker.benchmark', 3), ('benchmarking.InductorBenchmarker.benchmark_gpu', 3), ('fxgraph_cache_miss', 1), ('async_compile_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2783783Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2783990Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2784080Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2784287Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2784378Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2784461Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2784517Z Traceback (most recent call last): 2025-12-04T09:54:35.2784647Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2784687Z self.common( 2025-12-04T09:54:35.2784777Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2784823Z return func(*args, **kwds) 2025-12-04T09:54:35.2784951Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.2784996Z check_model( 2025-12-04T09:54:35.2785115Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2785167Z assert_equal_fn( 2025-12-04T09:54:35.2785307Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2785372Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2785533Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2785608Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2785665Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2785669Z 2025-12-04T09:54:35.2785714Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2785822Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2785927Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.2785931Z 2025-12-04T09:54:35.2785977Z The failure occurred for item [2] 2025-12-04T09:54:35.2785979Z 2025-12-04T09:54:35.2786054Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2786210Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2786223Z 2025-12-04T09:54:35.2786314Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2786394Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2786456Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2786559Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2786959Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2787038Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2787249Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2787338Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2787545Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2787635Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2787843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2787931Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2788139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2788224Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2788298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2788355Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2788454Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2788771Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2788847Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2789081Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2789169Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2789378Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2789465Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2789547Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2789598Z Traceback (most recent call last): 2025-12-04T09:54:35.2789724Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2789767Z self.common( 2025-12-04T09:54:35.2789856Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2789902Z return func(*args, **kwds) 2025-12-04T09:54:35.2790029Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.2790067Z check_model( 2025-12-04T09:54:35.2790206Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2790246Z assert_equal_fn( 2025-12-04T09:54:35.2790383Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2790445Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2790603Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2790688Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2790751Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2790755Z 2025-12-04T09:54:35.2790799Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2790902Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2791005Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.2791008Z 2025-12-04T09:54:35.2791054Z The failure occurred for item [2] 2025-12-04T09:54:35.2791056Z 2025-12-04T09:54:35.2791130Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2791289Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2791291Z 2025-12-04T09:54:35.2791379Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2791455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2791516Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2791616Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2791936Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2792014Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2792224Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2792311Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2792518Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2792617Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2792826Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2792913Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2793125Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2793212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2793286Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2793344Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2793445Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2793766Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2793841Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2794067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2794153Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2794360Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2794457Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2794539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2794602Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2794700Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2795024Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2795099Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2795309Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2795397Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2795606Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2795694Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2795903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2795991Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2796200Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2796286Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2796360Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2796419Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2796528Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2796885Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2796960Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2797167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2797257Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2797465Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2797551Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2797632Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2797679Z Traceback (most recent call last): 2025-12-04T09:54:35.2797834Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2797871Z self.common( 2025-12-04T09:54:35.2797958Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2798001Z return func(*args, **kwds) 2025-12-04T09:54:35.2798126Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.2798164Z check_model( 2025-12-04T09:54:35.2798280Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2798331Z assert_equal_fn( 2025-12-04T09:54:35.2798482Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2798544Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2798703Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2798776Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2798829Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2798831Z 2025-12-04T09:54:35.2798877Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2798978Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2799081Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.2799083Z 2025-12-04T09:54:35.2799127Z The failure occurred for item [2] 2025-12-04T09:54:35.2799131Z 2025-12-04T09:54:35.2799206Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2799359Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2799361Z 2025-12-04T09:54:35.2799449Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2799524Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2799584Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2799684Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2800004Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2800081Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2800310Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2800397Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2800607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2800694Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2800902Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2800988Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2801196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2801284Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2801357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2801434Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2801534Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2801857Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2801930Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2802157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2802245Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2802450Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2802539Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2802612Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2802673Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2802770Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2803089Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2803163Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2803371Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2803457Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2803663Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2803749Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2803958Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2804044Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2804262Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2804349Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2804422Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2804479Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2804576Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2804893Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2804967Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2805175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2805261Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2805478Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2805562Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2805636Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2805692Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2805790Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2806131Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2806207Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2806416Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2806504Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2806712Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2806832Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2806918Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2806964Z Traceback (most recent call last): 2025-12-04T09:54:35.2807095Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2807131Z self.common( 2025-12-04T09:54:35.2807219Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2807263Z return func(*args, **kwds) 2025-12-04T09:54:35.2807390Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.2807426Z check_model( 2025-12-04T09:54:35.2807544Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2807581Z assert_equal_fn( 2025-12-04T09:54:35.2807720Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2807781Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2807955Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2808028Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2808082Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2808085Z 2025-12-04T09:54:35.2808128Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2808225Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2808320Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.2808322Z 2025-12-04T09:54:35.2808367Z The failure occurred for item [2] 2025-12-04T09:54:35.2808369Z 2025-12-04T09:54:35.2808443Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2808599Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2808601Z 2025-12-04T09:54:35.2808691Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2808766Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2808826Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2808925Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2809258Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2809332Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2809555Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2809658Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2809867Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2809954Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2810163Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2810249Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2810456Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2810543Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2810618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2810674Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2810774Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2811091Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2811167Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2811374Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2811463Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2811671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2811769Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2811842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2811903Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2812001Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2812319Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2812394Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2812602Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2812689Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2812895Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2812994Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2813199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2813286Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2813510Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2813599Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2813671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2813729Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2813829Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2814147Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2814222Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2814432Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2814521Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2814726Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2814814Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2814887Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2814945Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2815042Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2815361Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2815451Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2815659Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2815744Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2815951Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2816037Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2816109Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2816169Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2816267Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2816588Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2816661Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2816914Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2817000Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2817210Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2817308Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2817527Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2817613Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2817820Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2817906Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2817987Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2818034Z Traceback (most recent call last): 2025-12-04T09:54:35.2818160Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2818196Z self.common( 2025-12-04T09:54:35.2818284Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2818328Z return func(*args, **kwds) 2025-12-04T09:54:35.2818454Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.2818490Z check_model( 2025-12-04T09:54:35.2818606Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2818644Z assert_equal_fn( 2025-12-04T09:54:35.2818782Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2818843Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2819000Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2819072Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2819125Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2819128Z 2025-12-04T09:54:35.2819174Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2819284Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2819378Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.2819381Z 2025-12-04T09:54:35.2819425Z The failure occurred for item [2] 2025-12-04T09:54:35.2819428Z 2025-12-04T09:54:35.2819501Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2819652Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2819654Z 2025-12-04T09:54:35.2819741Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2819815Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2819875Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2819976Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2820297Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2820385Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2820595Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2820682Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2820892Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2820998Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2821206Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2821293Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2821501Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2821588Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2821661Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2821718Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2821816Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2822142Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2822217Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2822429Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2822515Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2822723Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2822809Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2822882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2822954Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2823053Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2823372Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2823445Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2823655Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2823742Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2823952Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2824039Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2824245Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2824341Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2824547Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2824632Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2824707Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2824773Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2824882Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2825201Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2825275Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2825483Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2825568Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2825775Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2825860Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2825934Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2825990Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2826088Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2826407Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2826481Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2826688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2826825Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2827049Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2827135Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2827208Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2827268Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2827365Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2827685Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2827759Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2827968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2828054Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2828274Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2828361Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2828565Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2828650Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2828888Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2828976Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2829049Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2829110Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2829208Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2829530Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2829602Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2829813Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2829901Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2830108Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2830196Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2830404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2830490Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2830698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2830784Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2830879Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2830928Z Traceback (most recent call last): 2025-12-04T09:54:35.2831055Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2831093Z self.common( 2025-12-04T09:54:35.2831179Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2831223Z return func(*args, **kwds) 2025-12-04T09:54:35.2831349Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.2831385Z check_model( 2025-12-04T09:54:35.2831502Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2831540Z assert_equal_fn( 2025-12-04T09:54:35.2831679Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2831740Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2831899Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2831972Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2832037Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2832039Z 2025-12-04T09:54:35.2832086Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2832181Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2832277Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.2832279Z 2025-12-04T09:54:35.2832323Z The failure occurred for item [2] 2025-12-04T09:54:35.2832325Z 2025-12-04T09:54:35.2832409Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2832574Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2832576Z 2025-12-04T09:54:35.2832663Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2832738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2832797Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2832898Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2833218Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2833294Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2833505Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2833594Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2833802Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2833891Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2834097Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2834184Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2834391Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2834490Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2834563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2834621Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2834722Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2835041Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2835116Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2835324Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2835412Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2835618Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2835716Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2835788Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2835847Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2835944Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2836279Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2836364Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2836576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2836662Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2836905Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2836991Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2837199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2837286Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2837492Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2837585Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2837658Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2837717Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2837814Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2838137Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2838212Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2838442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2838528Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2838737Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2838823Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2838896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2838953Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2839051Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2839373Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2839447Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2839655Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2839755Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2839962Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2840046Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2840133Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2840207Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2840306Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2840623Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2840698Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2840905Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2840991Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2841197Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2841283Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2841491Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2841577Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2841784Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2841868Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2841941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2841999Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2842099Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2842434Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2842510Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2842718Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2842804Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2843010Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2843098Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2843304Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2843390Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2843599Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2843694Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2843768Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2843826Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2843924Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2844270Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2844345Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2844552Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2844639Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2844844Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2844930Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2845137Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2845225Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2845430Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2845517Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2845599Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2845647Z Traceback (most recent call last): 2025-12-04T09:54:35.2845772Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2845811Z self.common( 2025-12-04T09:54:35.2845899Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2845943Z return func(*args, **kwds) 2025-12-04T09:54:35.2846070Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.2846120Z check_model( 2025-12-04T09:54:35.2846236Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2846276Z assert_equal_fn( 2025-12-04T09:54:35.2846413Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2846474Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2846633Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2846705Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2846793Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2846795Z 2025-12-04T09:54:35.2846839Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2846937Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2847030Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.2847032Z 2025-12-04T09:54:35.2847077Z The failure occurred for item [2] 2025-12-04T09:54:35.2847079Z 2025-12-04T09:54:35.2847151Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2847321Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2847323Z 2025-12-04T09:54:35.2847409Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2847483Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2847541Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2847655Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2847988Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2848065Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2848284Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2848372Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2848579Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2848665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2848876Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2848962Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2849167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2849253Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2849327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2849383Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2849482Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2849804Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2849895Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2850102Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2850190Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2850398Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2850485Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2850556Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2850618Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2850717Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2851038Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2851122Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2851329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2851415Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2851631Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2851728Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2851935Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2852021Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2852227Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2852313Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2852385Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2852443Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2852543Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2852866Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2852940Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2853148Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2853235Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2853440Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2853527Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2853600Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2853673Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2853770Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2854088Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2854162Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2854371Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2854457Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2854666Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2854752Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2854825Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2854894Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2854993Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2855314Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2855387Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2855619Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2855706Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2855912Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2855998Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2856204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2856289Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2856497Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2856584Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2856657Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2856716Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2856854Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2857173Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2857247Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2857455Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2857559Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2857765Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2857851Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2858059Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2858144Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2858352Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2858438Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2858513Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2858572Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2858673Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2858991Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2859086Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2859296Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2859394Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2859614Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2859701Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2859907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2859993Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2860200Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2860285Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2860359Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2860418Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2860517Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2860834Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2860909Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2861117Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2861203Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2861411Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2861508Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2861715Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2861802Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2862007Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2862092Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2862174Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2862220Z Traceback (most recent call last): 2025-12-04T09:54:35.2862348Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2862386Z self.common( 2025-12-04T09:54:35.2862474Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2862518Z return func(*args, **kwds) 2025-12-04T09:54:35.2862645Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.2862695Z check_model( 2025-12-04T09:54:35.2862812Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2862849Z assert_equal_fn( 2025-12-04T09:54:35.2862987Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2863047Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2863204Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2863297Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2863355Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2863357Z 2025-12-04T09:54:35.2863400Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2863503Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2863605Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.2863608Z 2025-12-04T09:54:35.2863654Z The failure occurred for item [2] 2025-12-04T09:54:35.2863656Z 2025-12-04T09:54:35.2863729Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2866204Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2866208Z 2025-12-04T09:54:35.2866300Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2866378Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2866441Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2866547Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2866902Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2866980Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2867192Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2867281Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2867490Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2867604Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2867810Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2867897Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2868104Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2868189Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2868262Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2868320Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2868421Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2868747Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2868835Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2869044Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2869131Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2869347Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2869447Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2869522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2869582Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2869679Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2870003Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2870077Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2870283Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2870373Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2870584Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2870677Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2870882Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2870968Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2871173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2871259Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2871332Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2871400Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2871499Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2871846Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2871920Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2872128Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2872215Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2872424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2872512Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2872585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2872652Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2872750Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2873070Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2873143Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2873371Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2873458Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2873665Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2873751Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2873824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2873884Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2873982Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2874302Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2874377Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2874586Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2874673Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2874880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2874966Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2875175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2875262Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2875489Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2875574Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2875648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2875706Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2875805Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2876125Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2876201Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2876410Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2876497Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2876716Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2876839Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2877046Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2877132Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2877367Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2877454Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2877527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2877586Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2877686Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2878006Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2878080Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2878289Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2878376Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2878583Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2878669Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2878875Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2878962Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2879170Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2879270Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2879344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2879402Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2879501Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2879820Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2879893Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2880100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2880188Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2880393Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2880479Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2880697Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2880784Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2880989Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2881085Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2881167Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2881228Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2881326Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2881644Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2881719Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2881927Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2882014Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2882219Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2882307Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2882513Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2882600Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2882811Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2882897Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2882970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2883029Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2883136Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2883453Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2883529Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2883736Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2883822Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2884030Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2884116Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2884197Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2884245Z Traceback (most recent call last): 2025-12-04T09:54:35.2884372Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2884419Z self.common( 2025-12-04T09:54:35.2884508Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2884552Z return func(*args, **kwds) 2025-12-04T09:54:35.2884677Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.2884714Z check_model( 2025-12-04T09:54:35.2884831Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2884878Z assert_equal_fn( 2025-12-04T09:54:35.2885026Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2885089Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2885247Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2885321Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2885375Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2885377Z 2025-12-04T09:54:35.2885423Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2885520Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2885614Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.2885616Z 2025-12-04T09:54:35.2885661Z The failure occurred for item [2] 2025-12-04T09:54:35.2885663Z 2025-12-04T09:54:35.2885738Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2885892Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2885896Z 2025-12-04T09:54:35.2885984Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2886059Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2886118Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2886218Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2886536Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2886611Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2886858Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2886946Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2887153Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2887240Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2887446Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2887532Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2887739Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2887826Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2887898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2887955Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2888071Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2888390Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2888463Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2888701Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2888791Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2888997Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2889086Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2889159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2889218Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2889315Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2889640Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2889714Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2889921Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2890008Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2890216Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2890302Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2890512Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2890599Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2890817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2890902Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2890975Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2891032Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2891130Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2891448Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2891523Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2891732Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2891817Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2892034Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2892119Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2892192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2892249Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2892347Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2892690Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2892765Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2892975Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2893060Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2893266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2893350Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2893424Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2893484Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2893582Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2893901Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2893976Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2894182Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2894268Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2894476Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2894573Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2894779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2894866Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2895073Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2895157Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2895230Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2895289Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2895389Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2895707Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2895790Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2895996Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2896082Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2896288Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2896393Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2896600Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2896686Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2896919Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2897006Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2897079Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2897137Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2897235Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2897554Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2897628Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2897835Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2897920Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2898125Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2898212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2898418Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2898521Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2898727Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2898814Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2898886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2898945Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2899043Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2899367Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2899441Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2899649Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2899748Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2899953Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2900040Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2900258Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2900359Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2900565Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2900651Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2900724Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2900783Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2900881Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2901201Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2901276Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2901488Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2901573Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2901781Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2901867Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2902072Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2902159Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2902364Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2902462Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2902535Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2902594Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2902692Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2903013Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2903085Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2903294Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2903381Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2903591Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2903686Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2903760Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2903818Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2903916Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2904259Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2904333Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2904540Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2904626Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2904832Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2904916Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2905123Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2905209Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2905414Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2905499Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2905574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2905632Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2905731Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2906050Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2906140Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2906348Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2906433Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2906641Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2906726Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2906985Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2907071Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2907277Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2907362Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2907444Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2907506Z Traceback (most recent call last): 2025-12-04T09:54:35.2907634Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2907670Z self.common( 2025-12-04T09:54:35.2907760Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2907802Z return func(*args, **kwds) 2025-12-04T09:54:35.2907929Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.2907965Z check_model( 2025-12-04T09:54:35.2908109Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2908149Z assert_equal_fn( 2025-12-04T09:54:35.2908286Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2908347Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2908506Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2908577Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2908630Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2908633Z 2025-12-04T09:54:35.2908677Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2908781Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2908883Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.2908886Z 2025-12-04T09:54:35.2908933Z The failure occurred for item [2] 2025-12-04T09:54:35.2908935Z 2025-12-04T09:54:35.2909009Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2909161Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2909165Z 2025-12-04T09:54:35.2909254Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2909329Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2909390Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2909489Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2909810Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2909898Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2910108Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2910196Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2910403Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2910490Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2910695Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2910783Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2910988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2911074Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2911160Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2911217Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2911316Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2911636Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2911730Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2911940Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2912025Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2912234Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2912320Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2912393Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2912451Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2912549Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2912870Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2912945Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2913155Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2913239Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2913446Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2913531Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2913739Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2913837Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2914042Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2914129Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2914202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2914258Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2914358Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2914676Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2914751Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2914959Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2915056Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2915264Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2915349Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2915421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2915486Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2915601Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2915922Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2915996Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2916204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2916291Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2916497Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2916584Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2916657Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2916717Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2916856Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2917178Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2917251Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2917458Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2917546Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2917767Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2917853Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2918062Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2918148Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2918354Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2918441Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2918515Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2918574Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2918672Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2918990Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2919083Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2919291Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2919377Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2919620Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2919711Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2919917Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2920004Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2920212Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2920298Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2920370Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2920431Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2920529Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2920849Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2920923Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2921132Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2921218Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2921425Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2921512Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2921729Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2921815Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2922021Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2922106Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2922178Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2922237Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2922336Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2922655Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2922729Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2922950Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2923036Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2923242Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2923327Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2923551Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2923637Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2923843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2923929Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2924001Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2924059Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2924157Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2924479Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2924553Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2924762Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2924849Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2925055Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2925140Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2925350Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2925453Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2925660Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2925745Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2925820Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2925877Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2925975Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2926295Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2926369Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2926577Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2926662Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2926922Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2927008Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2927081Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2927139Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2927252Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2927583Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2927657Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2927864Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2927950Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2928159Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2928246Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2928454Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2928540Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2928746Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2928832Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2928907Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2928966Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2929066Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2929387Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2929480Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2929687Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2929774Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2929980Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2930065Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2930272Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2930359Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2930565Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2930665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2930738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2930796Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2930894Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2931224Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2931309Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2931517Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2931604Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2931809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2931895Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2932101Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2932188Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2932396Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2932482Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2932554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2932613Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2932710Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2933031Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2933105Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2933332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2933419Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2933625Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2933712Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2933793Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2933840Z Traceback (most recent call last): 2025-12-04T09:54:35.2933966Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2934004Z self.common( 2025-12-04T09:54:35.2934092Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2934136Z return func(*args, **kwds) 2025-12-04T09:54:35.2934261Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.2934297Z check_model( 2025-12-04T09:54:35.2934412Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2934464Z assert_equal_fn( 2025-12-04T09:54:35.2934600Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2934661Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2934819Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2934892Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2934958Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2934960Z 2025-12-04T09:54:35.2935017Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2935120Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2935224Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.2935226Z 2025-12-04T09:54:35.2935272Z The failure occurred for item [2] 2025-12-04T09:54:35.2935274Z 2025-12-04T09:54:35.2935348Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2935501Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2935503Z 2025-12-04T09:54:35.2935591Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2935666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2935726Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2935827Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2936149Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2936224Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2936436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2936524Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2936732Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2936861Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2937089Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2937176Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2937383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2937469Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2937542Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2937600Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2937701Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2938024Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2938100Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2938323Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2938410Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2938616Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2938702Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2938801Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2938862Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2938959Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2939278Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2939351Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2939559Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2939644Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2939852Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2939939Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2940144Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2940233Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2940438Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2940524Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2940597Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2940655Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2940769Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2941087Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2941161Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2941369Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2941454Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2941661Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2941747Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2941821Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2941878Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2941976Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2942306Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2942379Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2942599Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2942694Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2942903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2942988Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2943062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2943121Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2943219Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2943536Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2943612Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2943820Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2943906Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2944112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2944199Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2944405Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2944490Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2944698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2948176Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2948249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2948309Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2948408Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2948728Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2948802Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2949010Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2949097Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2949303Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2955691Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2955898Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2955983Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2956206Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2956316Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2956390Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2956447Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2956546Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2956898Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2956973Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2957182Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2957269Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2957474Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2957559Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2957764Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2957850Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2958054Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2958140Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2958213Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2958290Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2958388Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2958707Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2958782Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2958988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2959076Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2959284Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2959371Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2959577Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2959678Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2959883Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2959968Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2960042Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2960128Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2960227Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2960548Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2960622Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2960829Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2960915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2961121Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2961209Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2961415Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2961502Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2961708Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2961794Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2961866Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2961923Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2962022Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2962356Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2962430Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2962639Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2962725Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2962932Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2963019Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2963094Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2963152Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2963249Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2963570Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2963657Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2963864Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2963959Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2964178Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2964264Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2964469Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2964556Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2964763Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2964849Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2964925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2964983Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2965084Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2965403Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2965477Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2965687Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2965774Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2965984Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2966083Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2966291Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2966377Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2966584Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2966670Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2966790Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2966848Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2966950Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2967268Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2967359Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2967567Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2967652Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2967859Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2967959Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2968182Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2968268Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2968475Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2968560Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2968635Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2968693Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2968790Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2969110Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2969184Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2969391Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2969479Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2969684Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2969769Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2969843Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2969916Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2970017Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2970338Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2970412Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2970619Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2970706Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2970914Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2971003Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2971209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2971309Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2971516Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2971605Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2971677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2971736Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2971860Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2972183Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2972259Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2972470Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2972560Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2972768Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2972855Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2973061Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2973150Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2973359Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2973448Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2973522Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2973583Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2973680Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2974000Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2974085Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2974294Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2974388Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2974598Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2974687Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2974770Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.2974824Z Traceback (most recent call last): 2025-12-04T09:54:35.2974951Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.2974989Z self.common( 2025-12-04T09:54:35.2975076Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.2975130Z return func(*args, **kwds) 2025-12-04T09:54:35.2975256Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.2975293Z check_model( 2025-12-04T09:54:35.2975410Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.2975449Z assert_equal_fn( 2025-12-04T09:54:35.2975587Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.2975661Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.2975833Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.2975909Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.2975963Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.2975966Z 2025-12-04T09:54:35.2976015Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.2976116Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.2976221Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.2976223Z 2025-12-04T09:54:35.2976267Z The failure occurred for item [2] 2025-12-04T09:54:35.2976269Z 2025-12-04T09:54:35.2976342Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.2976497Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.2976499Z 2025-12-04T09:54:35.2976589Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.2976663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2976722Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2976856Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2977176Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2977250Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2977460Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2977568Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2977774Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2977860Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2978068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2978154Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2978360Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2978448Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2978521Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2978580Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2978679Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2978996Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2979084Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2979290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2979391Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2979610Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2979697Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2979769Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2979830Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2979928Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2980248Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2980322Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2980532Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2980618Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2980825Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2980911Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2981117Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2981202Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2981410Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2981508Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2981580Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2981636Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2981736Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2982053Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2982126Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2982335Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2982421Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2982629Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2982714Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2982799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2982855Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2982955Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2983280Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2983368Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2983576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2983661Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2983868Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2983953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2984026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2984084Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2984184Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2984502Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2984576Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2984783Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2984870Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2985076Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2985165Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2985371Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2985469Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2985674Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2985761Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2985834Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2985892Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2985990Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2986309Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2986383Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2986589Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2986693Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2986942Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2987028Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2987261Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2987349Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2987554Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2987641Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2987713Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2987772Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2987870Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2988190Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2988265Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2988472Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2988560Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2988764Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2988851Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2989057Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2989143Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2989366Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2989452Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2989525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2989585Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2989683Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2990002Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2990075Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2990285Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2990371Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2990575Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2990677Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2990882Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2990968Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2991196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2991284Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2991356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2991414Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2991513Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2991833Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2991907Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2992115Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2992202Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2992407Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2992494Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2992698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2992784Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2992988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2993075Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2993158Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2993216Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.2993314Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.2993634Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.2993707Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2993916Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2994002Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2994210Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2994295Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2994380Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2994438Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2994537Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2994856Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2994941Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2995160Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2995247Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2995453Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2995539Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2995746Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2995831Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2996038Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2996124Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2996197Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2996255Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2996355Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2996672Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2996787Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2996997Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2997101Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2997308Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2997394Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2997600Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2997684Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2997892Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2997977Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2998051Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2998109Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.2998208Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.2998539Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.2998614Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.2998821Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2998930Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2999139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2999224Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2999433Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.2999517Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2999724Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.2999810Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.2999884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.2999942Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3000041Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3000364Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3000440Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3000646Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3000733Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3000942Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3001046Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3001120Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3001180Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3001278Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3001598Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3001671Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3001879Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3001966Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3002171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3002271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3002478Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3002563Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3002778Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3002874Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3002947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3003006Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3003104Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3003427Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3003501Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3003709Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3003796Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3004003Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3004089Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3004298Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3004383Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3004589Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3004675Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3004749Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3004821Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3004918Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3005236Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3005311Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3005518Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3005604Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3005811Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3005899Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3005971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3006040Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3006138Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3006460Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3006532Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3006806Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3006894Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3007101Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3007187Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3007269Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3007316Z Traceback (most recent call last): 2025-12-04T09:54:35.3007444Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3007480Z self.common( 2025-12-04T09:54:35.3007569Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3007613Z return func(*args, **kwds) 2025-12-04T09:54:35.3007741Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.3007777Z check_model( 2025-12-04T09:54:35.3007894Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3007933Z assert_equal_fn( 2025-12-04T09:54:35.3008072Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3008132Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3008292Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3008365Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3008419Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3008421Z 2025-12-04T09:54:35.3008466Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3008584Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3008686Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.3008689Z 2025-12-04T09:54:35.3008734Z The failure occurred for item [2] 2025-12-04T09:54:35.3008737Z 2025-12-04T09:54:35.3008810Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3008965Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3008967Z 2025-12-04T09:54:35.3009055Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3009129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3009189Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3009289Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3009611Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3009684Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3009907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3009995Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3010202Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3010317Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3010535Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3010621Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3010828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3010915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3010990Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3011047Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3011147Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3011469Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3011543Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3011751Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3011838Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3012044Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3012129Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3012202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3012262Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3012371Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3012688Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3012762Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3012970Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3013056Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3013264Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3013350Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3013556Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3013652Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3013857Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3013942Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3014015Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3014071Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3014196Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3014514Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3014589Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3014795Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3014881Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3015085Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3015171Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3015249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3015305Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3015404Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3015726Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3015799Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3016005Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3016091Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3016310Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3016396Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3016468Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3016527Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3016624Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3016987Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3017061Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3017273Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3017359Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3017564Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3017665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3017870Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3017956Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3018186Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3018273Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3018346Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3018409Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3018511Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3018834Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3018906Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3019119Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3019206Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3019414Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3019501Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3019709Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3019795Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3020001Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3020088Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3020175Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3020235Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3020333Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3020653Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3020725Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3020933Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3021019Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3021226Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3021310Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3021527Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3021612Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3021818Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3021904Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3021988Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3022058Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3022157Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3022478Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3022556Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3022765Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3022851Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3023059Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3023145Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3023358Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3023445Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3023653Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3023738Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3023811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3023870Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3023969Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3024297Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3024370Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3024578Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3024665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3024873Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3024959Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3025166Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3025250Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3025470Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3025554Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3025629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3025688Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3025800Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3026131Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3026206Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3026414Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3026501Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3026708Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3026834Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3026911Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3026970Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3027069Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3027388Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3027462Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3027673Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3027759Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3027968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3028072Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3028279Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3028367Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3028580Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3028666Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3028744Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3028805Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3028907Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3029229Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3029317Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3029523Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3029609Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3029833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3029938Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3030150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3030235Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3030443Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3030530Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3030604Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3030681Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3030786Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3031109Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3031184Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3031393Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3031484Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3031697Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3031785Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3031995Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3032095Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3032305Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3032394Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3032468Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3032525Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3032628Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3032956Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3033031Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3033238Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3033335Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3033543Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3033628Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3033700Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3033771Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3033889Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3034208Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3034282Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3034489Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3034576Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3034783Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3034872Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3035079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3035164Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3035373Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3035464Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3035537Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3035600Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3035699Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3036041Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3036116Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3036330Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3036415Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3036622Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3036710Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3036964Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3037050Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3037256Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3037355Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3037428Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3037484Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3037584Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3037926Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3038001Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3038209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3038297Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3038507Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3038592Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3038665Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3038723Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3038825Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3039141Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3039218Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3039430Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3039516Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3039724Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3039828Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3039904Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3039968Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3040070Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3040392Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3040469Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3040681Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3040771Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3042268Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3042359Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3042587Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3042673Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3042883Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3042986Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3043063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3043123Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3043226Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3043564Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3043639Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3043853Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3043939Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3044171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3044259Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3044469Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3044557Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3044767Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3044855Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3044928Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3044988Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3045103Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3045425Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3045498Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3045705Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3045789Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3045996Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3046082Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3046164Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3046258Z Traceback (most recent call last): 2025-12-04T09:54:35.3046396Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3046433Z self.common( 2025-12-04T09:54:35.3046522Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3046565Z return func(*args, **kwds) 2025-12-04T09:54:35.3046693Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.3046729Z check_model( 2025-12-04T09:54:35.3046891Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3046943Z assert_equal_fn( 2025-12-04T09:54:35.3047088Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3047149Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3047311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3047384Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3047437Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3047440Z 2025-12-04T09:54:35.3047488Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3047592Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3047697Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.3047699Z 2025-12-04T09:54:35.3047743Z The failure occurred for item [2] 2025-12-04T09:54:35.3047746Z 2025-12-04T09:54:35.3047826Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3047981Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3047983Z 2025-12-04T09:54:35.3048072Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3048149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3048211Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3048310Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3048637Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3048713Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3048942Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3049033Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3049243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3049330Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3049542Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3049631Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3049839Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3049928Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3050024Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3050098Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3050208Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3050542Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3050616Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3050842Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3050931Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3051148Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3051234Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3051309Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3051370Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3051470Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3051793Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3051870Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3052087Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3052177Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3052394Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3052483Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3052699Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3052790Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3053014Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3053101Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3053176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3053232Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3053340Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3053660Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3053740Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3053949Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3054052Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3054272Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3054359Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3054433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3054489Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3054603Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3054922Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3055000Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3055213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3055299Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3055504Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3055594Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3055667Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3055729Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3055828Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3056151Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3056229Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3056437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3056525Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3059274Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3059403Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3059613Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3059702Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3059916Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3060002Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3060076Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3060137Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3060237Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3060578Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3060665Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3060874Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3060959Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3061186Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3061271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3061477Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3061563Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3061781Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3061865Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3061939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3061998Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3062101Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3062421Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3062495Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3062705Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3062790Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3062998Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3063084Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3063290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3063398Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3063606Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3063692Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3063765Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3063822Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3063921Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3064240Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3064315Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3064540Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3065942Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3066690Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3067007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3071307Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3071421Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3071671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3071771Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3071882Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3071959Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3072082Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3072466Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3072559Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3072802Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3072912Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3073155Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3073254Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3073492Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3073587Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3073894Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3073990Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3074082Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3074149Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3074273Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3074639Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3074730Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3074988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3075090Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3075405Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3075560Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3075649Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3075716Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3075830Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3076215Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3076301Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3076541Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3076643Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3076909Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3077008Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3077247Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3077350Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3077579Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3077681Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3077764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3077832Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3077947Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3078315Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3078421Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3078658Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3078759Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3078997Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3079098Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3079336Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3079438Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3079670Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3079773Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3079878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3079964Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3080074Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3080437Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3080540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3080781Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3080876Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3081083Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3081174Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3081382Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3081470Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3081676Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3081765Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3081838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3081898Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3081998Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3082318Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3082392Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3082603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3082708Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3082917Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3083007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3083079Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3083141Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3083239Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3083561Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3083636Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3083864Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3083982Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3084191Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3084277Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3084485Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3084587Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3084797Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3084884Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3084961Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3085020Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3085123Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3085447Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3085524Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3085736Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3085823Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3086035Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3086121Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3086331Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3086417Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3086626Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3086728Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3086842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3086901Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3087002Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3087319Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3087396Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3087608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3087696Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3087933Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3088037Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3088114Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3088173Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3088277Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3088609Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3088689Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3088899Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3088990Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3089196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3089286Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3089359Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3089421Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3089522Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3089846Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3089924Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3090133Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3090222Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3090430Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3090521Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3090748Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3090839Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3091049Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3091139Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3091213Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3091276Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3091375Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3091698Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3091792Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3092014Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3092104Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3092310Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3092400Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3092619Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3092710Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3092916Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3093006Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3093080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3093140Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3093239Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3093559Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3093634Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3093845Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3093932Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3094141Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3094227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3094304Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3094362Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3094466Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3094808Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3094884Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3095096Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3095183Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3095393Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3095480Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3095566Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3095618Z Traceback (most recent call last): 2025-12-04T09:54:35.3095776Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3095826Z self.common( 2025-12-04T09:54:35.3095923Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3095973Z return func(*args, **kwds) 2025-12-04T09:54:35.3096105Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.3096143Z check_model( 2025-12-04T09:54:35.3096265Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3096307Z assert_equal_fn( 2025-12-04T09:54:35.3096462Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3096528Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3096695Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3096908Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3096970Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3096974Z 2025-12-04T09:54:35.3097024Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3097129Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3097223Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.3097226Z 2025-12-04T09:54:35.3097278Z The failure occurred for item [2] 2025-12-04T09:54:35.3097281Z 2025-12-04T09:54:35.3097377Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3097540Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3097544Z 2025-12-04T09:54:35.3097639Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3097717Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3097784Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3097886Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3098211Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3098287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3098503Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3098613Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3098828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3098916Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3099127Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3099214Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3099426Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3099514Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3099591Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3099652Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3099769Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3100104Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3100179Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3100403Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3100495Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3100703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3100791Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3100869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3100929Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3101033Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3101350Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3101427Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3101634Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3101724Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3101935Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3102021Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3102234Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3102321Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3102530Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3102634Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3102711Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3102769Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3102871Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3103188Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3103267Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3103474Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3103565Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3103780Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3103879Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3103952Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3104012Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3104110Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3104447Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3104526Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3104734Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3104825Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3105030Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3105119Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3105193Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3105258Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3105357Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3105679Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3105755Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3105967Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3106054Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3106265Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3106367Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3106574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3106665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3106900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3106989Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3107062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3107125Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3107228Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3107549Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3107641Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3107873Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3107959Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3108169Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3108268Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3108479Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3108564Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3108774Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3108863Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3108937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3109001Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3109099Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3109419Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3109492Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3109702Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3109789Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3109996Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3110083Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3110292Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3110391Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3110602Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3110689Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3110765Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3110825Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3110928Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3111252Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3111327Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3111550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3111647Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3111854Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3111941Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3112171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3112275Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3114344Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3114435Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3114509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3114742Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3114841Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3115160Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3115239Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3115454Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3115541Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3115752Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3115844Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3116050Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3116141Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3116349Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3116460Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3116534Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3116595Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3116695Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3117057Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3117131Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3117409Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3117498Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3117745Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3117845Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3117922Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3117982Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3118084Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3118425Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3118502Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3118715Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3118803Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3119014Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3119100Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3119309Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3119396Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3119608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3119695Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3119774Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3119833Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3119936Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3120256Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3120335Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3120573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3120669Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3120879Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3120965Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3121173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3121258Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3121469Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3121554Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3121690Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3121759Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3121863Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3122181Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3122258Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3122479Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3122569Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3122779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3122867Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3123076Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3123161Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3123372Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3123458Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3123536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3123593Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3123700Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3124022Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3124099Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3124308Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3124396Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3124617Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3124707Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3124785Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3124847Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3124946Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3125267Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3125345Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3125553Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3125660Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3125876Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3125966Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3126172Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3126262Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3126480Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3126571Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3126646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3126708Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3126841Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3127162Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3127238Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3127445Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3127535Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3127741Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3127831Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3128037Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3128126Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3128332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3128438Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3128511Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3128571Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3128669Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3128988Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3129062Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3129273Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3129361Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3129570Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3129675Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3129762Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3129821Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3129920Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3130263Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3130337Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3130548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3130636Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3130845Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3130931Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3131007Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3131067Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3131170Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3131487Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3131569Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3131779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3131866Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3132078Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3132163Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3132372Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3132475Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3132683Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3132771Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3132846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3132905Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3133005Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3133323Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3133400Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3133623Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3133729Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3133935Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3134026Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3134250Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3134337Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3134549Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3134636Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3134713Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3134771Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3134873Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3135192Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3135271Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3135479Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3135570Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3135777Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3135866Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3135939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3136002Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3136101Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3136437Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3136518Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3136729Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3136852Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3137058Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3137149Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3137224Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3137288Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3137387Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3137727Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3137815Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3138031Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3138121Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3140301Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3140392Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3140607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3140702Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3140912Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3141006Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3141090Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3141153Z Traceback (most recent call last): 2025-12-04T09:54:35.3141286Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3141335Z self.common( 2025-12-04T09:54:35.3141427Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3141481Z return func(*args, **kwds) 2025-12-04T09:54:35.3141612Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.3141663Z check_model( 2025-12-04T09:54:35.3141782Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3141830Z assert_equal_fn( 2025-12-04T09:54:35.3141971Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3142044Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3142208Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3144873Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3144931Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3144933Z 2025-12-04T09:54:35.3144992Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3145094Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3145196Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.3145200Z 2025-12-04T09:54:35.3145250Z The failure occurred for item [2] 2025-12-04T09:54:35.3145253Z 2025-12-04T09:54:35.3145335Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3145492Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3145494Z 2025-12-04T09:54:35.3145591Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3145670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3145739Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3152699Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3153052Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3153136Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3153350Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3153464Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3153676Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3153772Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3153982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3154078Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3154287Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3154382Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3154459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3154527Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3154629Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3154958Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3155035Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3155251Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3155347Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3155556Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3155668Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3155745Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3155813Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3155917Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3156246Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3156323Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3156539Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3156628Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3156893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3156997Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3157213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3157301Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3157514Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3157622Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3157700Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3157765Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3157867Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3158192Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3158267Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3158482Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3158572Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3158786Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3158875Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3158956Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3159015Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3159120Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3159440Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3159523Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3159749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3159845Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3160060Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3160150Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3160233Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3160295Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3160402Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3160724Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3160808Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3161039Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3161145Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3161352Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3161447Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3161668Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3161765Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3161974Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3162069Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3162145Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3162212Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3162320Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3162641Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3162725Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3162936Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3163031Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3163240Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3163334Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3163542Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3163638Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3163859Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3163955Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3164032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3164098Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3164200Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3164523Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3164605Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3164815Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3164925Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3165143Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3165237Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3165446Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3165540Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3165763Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3165859Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3165937Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3166005Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3166107Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3166433Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3166509Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3166727Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3166859Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3167068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3167163Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3167372Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3167466Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3167676Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3167770Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3167865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3167931Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3168034Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3168365Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3168441Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3168659Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3168751Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3168968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3169077Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3169304Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3169399Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3169608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3169702Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3169790Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3169857Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3169957Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3170286Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3170363Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3170576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3170664Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3170877Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3170968Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3171050Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3171112Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3171218Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3171536Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3171617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3171833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3171938Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3172154Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3172243Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3172462Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3172550Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3172767Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3172857Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3172939Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3173000Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3173122Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3173452Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3173535Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3173755Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3173853Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3174066Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3174156Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3174373Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3174462Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3174676Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3174765Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3174847Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3174908Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3175015Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3175335Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3175416Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3175626Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3175721Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3175928Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3176046Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3176260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3176349Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3176564Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3176652Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3176734Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3176832Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3176941Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3177276Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3177373Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3177583Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3177677Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3177900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3177995Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3178073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3178140Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3178244Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3178570Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3178652Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3178867Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3178962Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3179170Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3179266Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3179477Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3179571Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3179778Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3179872Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3179962Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3180029Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3180129Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3180455Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3180532Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3180749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3180843Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3181053Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3181149Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3181377Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3181480Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3181688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3181781Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3181869Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3181934Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3182036Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3182364Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3182441Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3182658Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3182746Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3182961Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3183057Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3183131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3183194Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3183295Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3183621Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3183696Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3183908Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3184011Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3184227Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3184317Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3184399Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3184460Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3184565Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3184886Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3184968Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3185189Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3185293Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3185509Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3185598Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3185815Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3185917Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3186134Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3186223Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3186306Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3186369Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3186476Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3186844Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3186925Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3187135Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3187228Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3187437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3187529Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3187739Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3187825Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3188039Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3188146Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3188226Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3188284Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3188386Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3188706Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3188784Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3188994Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3189088Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3189308Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3189413Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3189487Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3189549Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3189650Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3189990Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3190071Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3190280Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3190373Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3190580Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3190669Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3190743Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3190808Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3190909Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3191234Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3191310Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3191524Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3191612Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3191823Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3191912Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3192133Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3192221Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3192431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3192523Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3192596Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3192659Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3192759Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3193078Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3193153Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3193375Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3193472Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3193687Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3193773Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3193994Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3194082Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3194294Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3194382Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3194468Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3194519Z Traceback (most recent call last): 2025-12-04T09:54:35.3194651Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3194691Z self.common( 2025-12-04T09:54:35.3194784Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3194832Z return func(*args, **kwds) 2025-12-04T09:54:35.3194963Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.3195008Z check_model( 2025-12-04T09:54:35.3195126Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3195171Z assert_equal_fn( 2025-12-04T09:54:35.3195311Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3195377Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3195538Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3195616Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3195673Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3195675Z 2025-12-04T09:54:35.3195726Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3195825Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3195933Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.3195935Z 2025-12-04T09:54:35.3195985Z The failure occurred for item [2] 2025-12-04T09:54:35.3195988Z 2025-12-04T09:54:35.3196066Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3196219Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3196221Z 2025-12-04T09:54:35.3196313Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3196387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3196448Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3196550Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3196903Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3197002Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3197228Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3197322Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3197529Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3197619Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3197840Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3197930Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3198136Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3198226Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3198299Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3198360Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3198460Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3198781Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3198855Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3199067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3199155Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3199363Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3199450Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3199527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3199587Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3199702Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3200021Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3200096Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3200306Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3200393Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3200603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3200691Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3200901Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3200998Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3201218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3201305Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3201382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3201439Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3201552Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3201871Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3201950Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3202161Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3202248Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3202455Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3202542Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3202618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3202675Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3202776Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3203094Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3203170Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3203378Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3203468Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3203677Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3203778Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3203852Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3203916Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3204015Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3204334Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3204410Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3204618Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3204709Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3204925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3205029Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3205236Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3205325Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3205541Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3205633Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3205707Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3205770Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3205870Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3206193Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3206266Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3206475Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3206566Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3206805Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3206896Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3207104Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3207196Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3207401Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3207492Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3207565Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3207653Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3207751Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3208072Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3208147Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3208360Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3208448Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3208659Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3208750Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3208968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3209071Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3209279Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3209368Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3209455Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3209517Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3209617Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3209940Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3210014Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3210226Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3210312Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3210522Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3210608Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3210818Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3210907Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3211115Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3211282Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3211357Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3211418Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3211518Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3211861Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3211936Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3212148Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3212234Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3212443Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3212530Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3212740Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3212826Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3213045Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3213142Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3213219Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3213280Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3213379Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3213715Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3213791Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3214002Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3214090Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3214298Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3214385Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3214463Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3214522Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3214624Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3214941Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3215019Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3215225Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3215316Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3215526Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3215627Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3215837Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3215925Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3216135Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3216221Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3216297Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3216355Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3216460Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3216830Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3216923Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3217129Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3217219Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3217424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3217531Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3217739Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3217830Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3218040Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3218126Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3218204Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3218262Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3218363Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3218680Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3218759Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3218966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3219065Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3219271Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3219362Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3219569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3219673Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3219880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3219971Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3220049Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3220108Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3220213Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3220536Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3220616Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3220849Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3220950Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3221156Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3221246Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3221319Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3221382Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3221494Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3221817Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3221893Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3222104Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3222191Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3222401Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3222492Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3222701Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3222793Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3223005Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3223096Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3223170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3223233Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3223332Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3223652Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3223741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3223953Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3224040Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3224251Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3224337Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3224548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3224639Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3224859Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3224962Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3225037Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3225099Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3225197Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3225533Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3225610Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3225822Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3225909Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3226120Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3226207Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3226285Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3226343Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3226446Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3226808Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3226889Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3227100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3227187Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3227400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3227486Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3227585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3227644Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3227748Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3228069Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3228148Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3228358Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3228449Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3228658Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3228761Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3228981Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3229073Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3229283Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3229370Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3229461Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3229523Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3229627Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3229946Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3230026Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3230235Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3230328Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3230536Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3230628Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3230837Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3230928Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3231137Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3231228Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3231303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3231365Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3231476Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3231800Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3231878Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3232086Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3232177Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3232383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3232474Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3232549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3232610Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3232722Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3233057Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3233131Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3233355Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3233444Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3233657Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3233750Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3233824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3233887Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3233988Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3234310Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3234384Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3234599Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3234687Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3234899Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3234986Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3235196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3235283Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3235495Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3235604Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3235682Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3235742Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3235846Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3236170Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3236245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3236458Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3236545Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3236805Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3236904Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3237116Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3237204Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3237437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3237525Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3237605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3237664Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3237772Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3238094Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3238173Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3238388Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3238476Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3238691Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3238779Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3238991Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3239078Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3239290Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3239377Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3239465Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3239539Z Traceback (most recent call last): 2025-12-04T09:54:35.3239671Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3239715Z self.common( 2025-12-04T09:54:35.3239810Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3239855Z return func(*args, **kwds) 2025-12-04T09:54:35.3239986Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.3240026Z check_model( 2025-12-04T09:54:35.3240149Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3240188Z assert_equal_fn( 2025-12-04T09:54:35.3240331Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3240393Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3240557Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3240631Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3240704Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3240717Z 2025-12-04T09:54:35.3240765Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3240875Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3240980Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.3240982Z 2025-12-04T09:54:35.3241034Z The failure occurred for item [2] 2025-12-04T09:54:35.3241036Z 2025-12-04T09:54:35.3241110Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3241285Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3241290Z 2025-12-04T09:54:35.3241383Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3241458Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3241524Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3241626Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3241946Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3242021Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3242236Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3242326Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3242538Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3242626Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3242836Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3242923Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3243132Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3243219Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3243314Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3243376Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3243478Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3243800Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3243876Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3244087Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3244175Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3244386Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3244486Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3244573Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3244633Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3244736Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3245053Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3245142Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3245351Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3245442Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3245655Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3245743Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3245954Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3246040Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3246250Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3246337Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3246416Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3246474Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3246577Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3246926Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3247004Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3247212Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3247319Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3247529Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3247621Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3247695Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3247756Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3247857Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3248182Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3248261Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3248484Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3248588Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3248795Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3248886Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3248960Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3249023Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3249141Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3249467Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3249542Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3249755Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3249842Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3250052Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3250144Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3250356Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3250447Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3250654Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3250744Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3250817Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3250881Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3250980Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3251302Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3251393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3251607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3251694Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3251906Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3251993Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3252205Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3252293Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3252520Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3252622Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3252696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3252759Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3252858Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3253192Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3253267Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3253480Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3253568Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3253779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3253865Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3254075Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3254161Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3254372Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3254459Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3254540Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3254599Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3254703Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3255027Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3255102Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3255329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3255416Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3255630Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3255716Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3255926Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3256012Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3256224Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3256312Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3256402Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3256472Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3256575Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3256930Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3257009Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3257237Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3257327Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3257539Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3257627Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3257838Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3257925Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3258139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3258227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3258305Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3258363Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3258467Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3258788Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3258866Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3259081Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3259190Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3259404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3259492Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3259571Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3259630Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3259733Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3260053Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3260133Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3260341Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3260454Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3260677Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3260766Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3260973Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3261076Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3261283Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3261376Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3261451Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3261515Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3261615Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3261935Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3262014Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3262220Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3262312Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3262520Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3262610Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3262816Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3262907Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3263113Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3263215Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3263289Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3263352Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3263452Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3263774Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3263853Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3264062Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3264154Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3264370Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3264476Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3264683Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3264773Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3264981Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3265084Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3265159Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3265221Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3265323Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3265646Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3265720Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3265933Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3266020Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3266233Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3266391Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3272753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3272823Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3272927Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3273248Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3273331Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3273585Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3273677Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3273887Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3273978Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3274184Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3274273Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3274485Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3274575Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3274654Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3274731Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3274850Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3275168Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3275245Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3275472Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3275565Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3275772Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3275862Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3276068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3276158Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3276363Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3276453Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3276527Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3276590Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3276690Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3277058Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3277136Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3277344Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3277435Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3277662Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3277753Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3277827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3277888Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3277986Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3278303Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3278376Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3278585Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3278671Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3278892Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3278995Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3279069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3279128Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3279226Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3279562Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3279635Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3279845Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3279931Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3280139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3280223Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3280431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3280519Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3280727Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3280813Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3280886Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3280944Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3281044Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3281362Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3281450Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3281659Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3281744Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3281950Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3282034Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3282241Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3282327Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3282533Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3282629Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3282714Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3282771Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3282870Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3283188Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3287875Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3288880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3289002Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3289229Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3289321Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3289419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3289481Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3289588Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3289924Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3290007Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3290220Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3290313Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3290520Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3290610Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3290686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3290949Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3291051Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3291380Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3291459Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3291670Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3291761Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3291968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3292058Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3292328Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3292461Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3292666Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3292755Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3292828Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3292891Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3293016Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3293367Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3293443Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3293654Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3293740Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3293949Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3294039Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3294244Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3294335Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3294542Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3294630Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3294703Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3294764Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3294863Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3295185Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3295271Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3295480Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3295566Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3295773Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3295860Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3296069Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3296158Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3296375Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3296475Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3296547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3296608Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3296706Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3297082Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3297157Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3297367Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3297452Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3297659Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3297745Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3297829Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3297885Z Traceback (most recent call last): 2025-12-04T09:54:35.3298023Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3298064Z self.common( 2025-12-04T09:54:35.3298162Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3298211Z return func(*args, **kwds) 2025-12-04T09:54:35.3298343Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.3298384Z check_model( 2025-12-04T09:54:35.3298504Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3298544Z assert_equal_fn( 2025-12-04T09:54:35.3298689Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3298751Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3298916Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3299004Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3299062Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3299067Z 2025-12-04T09:54:35.3299119Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3299252Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3299360Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.3299363Z 2025-12-04T09:54:35.3299412Z The failure occurred for item [2] 2025-12-04T09:54:35.3299415Z 2025-12-04T09:54:35.3299493Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3299648Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3299651Z 2025-12-04T09:54:35.3299744Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3299820Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3299885Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3299986Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3300328Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3300419Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3300635Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3300755Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3300967Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3301055Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3301266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3301356Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3301561Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3301649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3301723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3301783Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3301884Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3302207Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3302282Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3302491Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3302578Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3302787Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3302888Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3302963Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3303024Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3303127Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3303448Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3303525Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3303736Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3303823Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3304065Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3304162Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3304369Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3304454Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3304662Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3304761Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3304838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3304895Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3304997Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3305316Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3305391Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3305600Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3305689Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3305897Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3305983Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3306059Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3306118Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3306218Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3306536Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3306613Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3306858Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3306973Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3307180Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3307269Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3307341Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3307403Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3307502Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3307827Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3307902Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3308127Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3308232Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3308437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3308526Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3308764Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3308853Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3309059Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3309149Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3309221Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3309283Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3309381Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3309707Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3309781Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3309993Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3310080Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3310288Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3310376Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3310582Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3310670Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3310893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3310981Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3311055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3311116Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3311214Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3311533Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3311607Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3311818Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3311905Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3312126Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3312222Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3312432Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3312520Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3312739Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3312829Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3312902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3312964Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3313064Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3313386Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3313460Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3313671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3313758Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3313967Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3314054Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3314265Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3314351Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3314562Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3314649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3314734Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3314795Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3314895Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3315216Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3315291Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3315500Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3315587Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3315795Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3315893Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3316116Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3316202Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3316410Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3316496Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3316583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3316641Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3316955Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3317275Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3317351Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3317580Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3317667Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3317876Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3317962Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3318038Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3318096Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3318198Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3318515Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3318591Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3318799Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3318928Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3319136Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3319226Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3319433Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3319521Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3319731Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3319816Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3319893Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3319951Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3320064Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3320399Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3320476Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3320696Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3320785Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3320990Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3321079Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3321285Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3321373Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3321580Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3321668Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3321744Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3321803Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3321904Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3322222Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3322300Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3322505Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3322593Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3322798Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3322898Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3323104Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3323194Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3323400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3323488Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3323562Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3323623Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3323766Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3324096Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3324185Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3324392Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3324483Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3324702Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3324790Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3324867Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3324926Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3325028Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3325349Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3325425Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3325667Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3325759Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3325968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3326059Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3326266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3326355Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3326564Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3326654Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3326732Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3326847Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3326950Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3327271Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3327348Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3327555Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3327646Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3327853Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3327944Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3328176Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3328279Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3328486Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3328576Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3328649Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3328730Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3328829Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3329151Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3329230Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3329437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3329527Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3329735Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3329826Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3329899Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3329959Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3330059Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3330388Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3330462Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3330675Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3330785Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3330993Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3331080Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3331157Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3331215Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3331317Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3331640Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3331740Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3331950Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3332053Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3332274Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3332359Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3332569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3332679Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3332893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3332980Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3333058Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3333119Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3333220Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3333538Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3333617Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3333828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3333917Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3334128Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3334215Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3334425Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3334511Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3334721Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3334818Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3334896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3334954Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3335057Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3335377Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3335453Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3335660Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3335751Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3336028Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3336127Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3336203Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3336260Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3336364Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3336715Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3336823Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3337032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3337122Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3337332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3337423Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3337495Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3337559Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3337658Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3337979Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3338056Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3338267Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3338355Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3338560Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3338651Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3338856Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3338964Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3339171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3339261Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3339334Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3339397Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3339495Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3339820Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3339895Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3340117Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3340218Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3340427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3340516Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3340736Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3340827Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3341034Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3341126Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3341199Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3341261Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3341359Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3341683Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3341761Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3341976Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3342063Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3342274Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3342361Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3342573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3342660Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3342883Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3342974Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3343047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3343109Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3343207Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3343527Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3343602Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3343813Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3343899Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3344121Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3344219Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3344297Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3344354Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3344456Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3344786Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3344864Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3345077Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3345165Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3345375Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3345464Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3345550Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3345600Z Traceback (most recent call last): 2025-12-04T09:54:35.3345735Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3345777Z self.common( 2025-12-04T09:54:35.3345872Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3345918Z return func(*args, **kwds) 2025-12-04T09:54:35.3346049Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.3346087Z check_model( 2025-12-04T09:54:35.3346207Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3346250Z assert_equal_fn( 2025-12-04T09:54:35.3346392Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3346454Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3346620Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3346707Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3346799Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3346803Z 2025-12-04T09:54:35.3346853Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3346961Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3347065Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.3347067Z 2025-12-04T09:54:35.3347117Z The failure occurred for item [2] 2025-12-04T09:54:35.3347119Z 2025-12-04T09:54:35.3347194Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3347352Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3347356Z 2025-12-04T09:54:35.3347445Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3347525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3347586Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3347715Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3348058Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3348132Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3348368Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3348457Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3348669Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3348758Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3348969Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3349057Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3349267Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3349353Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3349431Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3349489Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3349593Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3349913Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3349990Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3350201Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3350289Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3350498Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3350603Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3350681Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3350742Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3350844Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3351163Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3351240Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3351449Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3351540Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3351759Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3351861Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3352066Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3352157Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3352373Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3352465Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3352543Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3352600Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3352705Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3353026Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3353103Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3353312Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3353402Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3353611Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3353701Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3353775Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3353836Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3353935Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3354259Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3354333Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3354559Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3354650Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3354857Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3354947Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3355020Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3355082Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3355182Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3355503Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3355595Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3355817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3355903Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3356112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3356197Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3356418Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3356506Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3356717Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3356849Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3356926Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3356988Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3357087Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3357407Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3357481Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3357710Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3357798Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3358009Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3358096Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3358305Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3358414Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3358623Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3358711Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3358787Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3358847Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3358948Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3359267Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3359344Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3359555Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3359665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3359889Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3359975Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3360185Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3360287Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3360497Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3360585Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3360663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3360723Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3360826Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3361145Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3361223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3361431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3361521Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3361734Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3361821Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3362030Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3362116Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3362326Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3362431Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3362507Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3362567Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3362670Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3362988Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3363064Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3363272Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3363363Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3363600Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3363707Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3363916Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3364004Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3364213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3364314Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3364392Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3364449Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3364551Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3364874Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3364952Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3365159Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3365249Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3365457Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3365547Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3365622Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3365685Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3365784Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3366107Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3366185Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3366393Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3366499Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3366707Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3366834Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3367040Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3367130Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3367337Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3367428Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3367501Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3367586Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3367697Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3368017Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3368091Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3368314Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3368405Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3368613Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3368702Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3368913Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3369002Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3369209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3369300Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3369374Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3369436Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3369535Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3369856Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3369930Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3370140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3370228Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3370482Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3370576Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3370783Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3370877Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3371085Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3371180Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3371255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3371321Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3371422Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3371769Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3371857Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3372073Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3372162Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3372388Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3372477Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3372557Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3372619Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3372726Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3373050Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3373163Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3373380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3373469Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3373682Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3373770Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3373984Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3374072Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3374285Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3374374Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3374474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3374536Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3374643Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3374964Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3375045Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3375260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3375350Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3375567Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3375672Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3375895Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3375984Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3376200Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3376290Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3376387Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3376448Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3376554Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3376902Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3376984Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3377193Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3377287Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3377494Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3377589Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3377670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3377731Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3377838Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3378156Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3378238Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3378449Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3378576Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3378785Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3378881Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3378958Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3379024Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3379124Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3379451Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3379527Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3379762Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3379863Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3380077Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3380170Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3380378Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3380487Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3380695Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3380788Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3380865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3380930Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3381030Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3381350Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3381425Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3381638Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3381725Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3381935Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3382024Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3382237Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3382330Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3382540Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3382651Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3382728Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3382794Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3382895Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3383224Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3383301Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3383518Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3383607Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3383835Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3383933Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3384014Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3384073Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3384180Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3384515Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3384600Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3384815Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3384904Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3385115Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3385202Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3385278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3385338Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3385440Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3385762Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3385842Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3386049Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3386138Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3386344Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3386433Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3386655Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3386782Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3386992Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3387077Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3387153Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3387211Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3387313Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3387633Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3387732Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3387951Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3388040Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3388248Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3388337Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3388565Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3388736Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3388942Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3389031Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3389104Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3389166Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3389267Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3389587Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3389665Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3389873Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3389966Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3390171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3390264Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3390469Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3390577Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3390782Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3390873Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3390947Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3391007Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3391105Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3391430Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3391506Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3391744Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3391846Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3392062Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3392153Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3392226Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3392332Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3402252Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3402574Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3402651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3402864Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3402950Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3403161Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3403248Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3403325Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3403385Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3403486Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3409490Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3409565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3409775Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3409862Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3410072Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3417031Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3417239Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3417326Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3417533Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3417621Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3417698Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3417757Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3417859Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3418199Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3418292Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3418501Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3418586Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3418814Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3418902Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3418986Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3419037Z Traceback (most recent call last): 2025-12-04T09:54:35.3419169Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3419210Z self.common( 2025-12-04T09:54:35.3419304Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3419349Z return func(*args, **kwds) 2025-12-04T09:54:35.3419478Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.3419517Z check_model( 2025-12-04T09:54:35.3419636Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3419679Z assert_equal_fn( 2025-12-04T09:54:35.3419822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3419887Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3420049Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3420124Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3420181Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3420183Z 2025-12-04T09:54:35.3420231Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3420337Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3420440Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.3420442Z 2025-12-04T09:54:35.3420490Z The failure occurred for item [2] 2025-12-04T09:54:35.3420493Z 2025-12-04T09:54:35.3420571Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3420782Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3420784Z 2025-12-04T09:54:35.3420873Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3420951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3421011Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3421113Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3421435Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3421509Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3421723Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3421812Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3422064Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3422172Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3422379Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3422465Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3422688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3422776Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3422851Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3422910Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3423013Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3423331Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3423407Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3423617Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3423704Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3423914Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3424001Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3424077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3424136Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3424237Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3424556Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3424649Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3424856Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3424945Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3425151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3425239Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3425445Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3425535Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3425743Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3425844Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3425931Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3425990Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3426090Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3426409Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3426503Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3426710Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3426871Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3427077Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3427166Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3427239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3427299Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3427396Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3427716Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3427790Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3428001Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3428091Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3428296Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3428385Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3428459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3428543Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3428641Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3428963Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3429037Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3429246Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3429334Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3429542Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3429629Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3429858Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3429965Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3430173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3430260Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3430333Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3430394Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3430506Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3430831Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3430905Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3431114Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3431200Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3431407Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3431493Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3431702Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3431788Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3431996Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3432082Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3432157Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3432215Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3432316Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3432637Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3432729Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3432941Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3433026Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3433234Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3433320Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3433530Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3433617Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3433840Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3433936Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3434011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3434071Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3434172Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3434502Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3434579Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3434789Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3434876Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3435086Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3435172Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3435380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3435466Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3435674Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3435760Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3435836Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3435894Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3435995Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3436314Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3436389Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3436618Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3436707Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3436956Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3437045Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3437251Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3437339Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3437548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3437634Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3437728Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3437801Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3437902Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3438220Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3438296Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3438516Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3438605Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3438811Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3438899Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3438971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3439033Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3439142Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3439463Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3439540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3439751Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3439842Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3440047Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3440134Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3440339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3440442Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3440647Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3440738Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3440811Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3440871Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3440968Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3441287Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3441360Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3441570Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3441674Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3441890Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3441977Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3442182Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3442278Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3442483Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3442570Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3442643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3442704Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3442801Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3443119Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3443193Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3443400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3443486Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3443694Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3443781Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3443985Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3444074Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3444280Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3444380Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3444453Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3444512Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3444610Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3444931Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3445004Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3445212Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3445298Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3445531Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3445628Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3445702Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3445760Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3445858Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3446189Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3446264Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3446474Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3446560Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3446808Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3446892Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3447100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3447185Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3447392Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3447477Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3447552Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3447611Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3447711Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3448026Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3448103Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3448311Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3448420Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3448629Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3448715Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3448922Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3449007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3449214Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3449300Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3449374Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3449443Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3449560Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3449876Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3449951Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3450168Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3450258Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3450464Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3450552Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3450627Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3450685Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3450784Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3451101Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3451177Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3451384Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3451472Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3451678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3451765Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3451837Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3451899Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3451996Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3452329Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3452403Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3452613Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3452700Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3452907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3452996Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3453202Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3453288Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3453503Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3453600Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3453671Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3453731Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3453829Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3454159Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3454234Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3454444Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3454530Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3454738Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3454823Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3455032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3455120Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3455327Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3455415Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3455487Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3455545Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3455642Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3455962Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3456052Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3456261Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3456351Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3456558Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3456643Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3456717Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3456812Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3456911Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3457251Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3457340Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3457551Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3457636Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3457844Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3457943Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3458020Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3458078Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3458179Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3458498Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3458574Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3458780Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3458868Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3459074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3459161Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3459366Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3459454Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3459662Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3459749Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3459824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3459905Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3460005Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3460322Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3460398Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3460606Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3460695Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3460900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3460988Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3461207Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3461303Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3461507Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3461595Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3461670Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3461728Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3461844Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3462165Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3462241Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3462447Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3462539Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3462745Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3462837Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3463045Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3463140Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3463349Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3463441Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3463516Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3463580Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3463679Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3464004Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3464103Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3464314Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3464408Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3464613Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3464704Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3464780Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3464845Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3464944Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3465276Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3465361Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3465574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3465662Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3465886Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3465975Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3466055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3466116Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3466219Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3466543Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3466618Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3466867Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3466955Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3467168Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3467256Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3467467Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3467555Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3467767Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3467854Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3467951Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3468009Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3468113Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3468433Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3468515Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3468733Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3468822Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3469034Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3469134Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3469225Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3469283Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3469388Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3469707Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3469803Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3470014Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3470107Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3470313Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3470406Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3470489Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3470548Z Traceback (most recent call last): 2025-12-04T09:54:35.3470677Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3470722Z self.common( 2025-12-04T09:54:35.3470813Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3470866Z return func(*args, **kwds) 2025-12-04T09:54:35.3470995Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.3471040Z check_model( 2025-12-04T09:54:35.3471159Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3471207Z assert_equal_fn( 2025-12-04T09:54:35.3471347Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3471414Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3471574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3471654Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3471714Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3471729Z 2025-12-04T09:54:35.3471779Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3471888Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3471993Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.3471996Z 2025-12-04T09:54:35.3472050Z The failure occurred for item [2] 2025-12-04T09:54:35.3472052Z 2025-12-04T09:54:35.3472127Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3472286Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3472288Z 2025-12-04T09:54:35.3472377Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3472456Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3472518Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3472625Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3472956Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3473045Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3473258Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3473353Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3473579Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3473673Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3473885Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3473974Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3474188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3474277Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3474356Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3474415Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3474521Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3474842Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3474923Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3475134Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3475227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3475433Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3475526Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3475601Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3475680Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3475781Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3476110Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3476191Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3476399Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3476492Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3476699Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3476830Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3477067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3477176Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3477381Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3477475Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3477549Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3477627Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3477729Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3478059Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3478135Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3478349Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3478441Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3478648Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3478743Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3478818Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3478882Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3478982Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3479305Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3479380Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3479594Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3479698Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3479910Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3479998Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3480077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3480139Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3480243Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3480566Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3480647Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3480860Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3480962Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3481184Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3481272Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3481483Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3481585Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3481798Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3481885Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3481968Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3482030Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3482135Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3482454Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3482535Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3482748Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3482840Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3483054Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3483142Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3483353Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3483440Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3483650Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3483761Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3483839Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3483900Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3484004Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3484325Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3484404Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3484613Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3484705Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3484949Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3485051Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3485262Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3485358Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3485566Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3485663Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3485740Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3485799Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3485901Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3486220Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3492389Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3492635Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3492728Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3492939Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3493024Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3493238Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3493328Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3493533Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3493621Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3493697Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3493852Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3493952Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3494276Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3494352Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3494561Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3494646Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3494856Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3494943Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3495165Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3495273Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3495481Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3495569Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3495644Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3495703Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3495824Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3496149Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3496223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3496432Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3496518Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3496726Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3496853Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3496930Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3496989Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3497090Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3497411Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3497486Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3497693Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3497781Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3498013Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3498099Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3498307Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3498392Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3498599Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3498685Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3498761Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3498820Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3498921Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3499256Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3499344Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3499550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3499640Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3499859Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3499949Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3500159Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3500245Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3500454Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3500539Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3500614Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3500673Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3500774Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3501094Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3501171Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3501380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3501468Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3501676Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3501765Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3501987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3502077Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3502285Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3502373Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3502451Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3502508Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3502612Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3502931Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3503020Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3503238Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3503325Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3503530Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3503615Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3503700Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3503761Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3503858Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3504177Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3504250Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3504459Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3504546Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3504753Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3504840Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3505046Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3505133Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3505339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3505425Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3505497Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3505558Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3505669Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3505987Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3506060Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3506267Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3506353Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3506559Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3506644Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3506888Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3506996Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3507218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3507304Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3507377Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3507435Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3507554Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3507874Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3507948Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3508158Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3508242Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3508449Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3508534Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3508609Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3508666Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3508764Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3509084Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3509160Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3509368Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3509454Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3509661Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3509764Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3509838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3509897Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3509996Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3510313Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3510386Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3510593Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3510681Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3510901Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3510999Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3511206Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3511293Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3511512Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3511599Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3511672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3511730Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3511832Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3512151Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3512225Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3512432Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3512520Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3512725Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3512814Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3513020Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3513107Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3513312Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3513400Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3513483Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3513557Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3513655Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3513972Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3514046Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3514252Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3514339Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3514548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3514635Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3514721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3514789Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3514886Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3515203Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3515276Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3515494Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3515580Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3515790Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3515877Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3515948Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3516007Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3516104Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3516422Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3516497Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3516705Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3516831Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3517036Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3517120Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3517327Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3517431Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3517638Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3517723Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3517799Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3517857Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3517958Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3518279Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3518354Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3518563Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3518664Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3518883Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3518968Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3519174Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3519305Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3519511Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3519597Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3519672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3519731Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3519829Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3520149Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3520224Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3520432Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3520518Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3520725Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3520811Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3521016Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3521100Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3521306Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3521405Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3521478Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3521536Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3521635Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3521951Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3522025Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3522232Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3522319Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3522539Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3522634Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3522707Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3522762Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3522860Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3523188Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3523263Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3523470Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3523557Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3523761Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3523849Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3523920Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3523980Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3524078Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3524425Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3524499Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3524709Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3524796Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3525000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3525087Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3525311Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3525398Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3525604Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3525691Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3525765Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3525823Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3525920Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3526239Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3526313Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3526539Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3526634Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3526900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3526987Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3527077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3527135Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3527234Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3527551Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3527624Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3527831Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3527914Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3528121Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3528206Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3528278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3528335Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3528433Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3528750Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3528823Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3529031Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3529134Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3529343Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3529429Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3529514Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3529562Z Traceback (most recent call last): 2025-12-04T09:54:35.3529694Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3529735Z self.common( 2025-12-04T09:54:35.3529827Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3529876Z return func(*args, **kwds) 2025-12-04T09:54:35.3530007Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.3530046Z check_model( 2025-12-04T09:54:35.3530163Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3530201Z assert_equal_fn( 2025-12-04T09:54:35.3530357Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3530430Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3530593Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3530666Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3530720Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3530723Z 2025-12-04T09:54:35.3530770Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3530885Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3530989Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.3530991Z 2025-12-04T09:54:35.3531037Z The failure occurred for item [2] 2025-12-04T09:54:35.3531039Z 2025-12-04T09:54:35.3531113Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3531268Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3531270Z 2025-12-04T09:54:35.3531359Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3531436Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3531494Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3531595Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3531916Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3531992Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3532203Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3532290Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3532500Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3532586Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3532792Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3532892Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3533099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3533185Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3533259Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3533314Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3533414Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3533735Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3533810Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3534029Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3534128Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3534334Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3534419Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3534491Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3534560Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3534659Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3534977Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3535052Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3535259Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3535345Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3535550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3535637Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3535842Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3535928Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3536135Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3536220Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3536291Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3536348Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3536445Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3536803Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3536892Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3537108Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3537196Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3537403Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3537488Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3537563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3537620Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3537719Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3538050Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3538138Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3538348Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3538434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3538657Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3538745Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3538818Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3538878Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3538976Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3539300Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3539374Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3539581Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3539669Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3539874Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3539961Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3540165Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3540251Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3540457Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3540543Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3540629Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3540689Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3540788Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3541110Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3541183Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3541393Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3541480Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3541687Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3541787Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3542001Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3542088Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3542292Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3542378Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3542459Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3542520Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3542618Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3542939Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3543012Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3543222Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3543307Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3543514Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3543602Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3543809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3543896Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3544100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3544215Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3544287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3544347Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3544456Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3544776Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3544852Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3545060Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3545145Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3545352Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3545438Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3545644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3545740Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3545955Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3546042Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3546114Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3546174Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3546281Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3546600Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3546673Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3546930Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3547016Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3547223Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3547309Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3547516Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3547603Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3547809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3547894Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3547970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3548026Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3548126Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3548446Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3548542Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3548751Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3548837Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3549044Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3549129Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3549203Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3549262Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3549362Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3549698Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3549785Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3549990Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3550078Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3550302Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3550389Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3550597Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3550682Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3550888Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3550973Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3551047Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3551104Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3551205Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3551526Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3551601Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3551808Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3551895Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3552100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3552187Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3552406Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3552493Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3552700Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3552788Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3552861Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3552919Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3553018Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3553338Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3553415Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3553660Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3553766Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3553978Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3554065Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3554281Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3554369Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3554576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3554663Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3554735Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3554793Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3554893Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3555211Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3555287Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3555495Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3555582Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3555787Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3555875Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3555948Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3556008Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3556106Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3556441Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3556514Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3556723Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3556849Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3557054Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3557142Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3557347Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3557434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3557655Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3557754Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3557827Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3557886Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3557984Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3558319Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3558394Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3558604Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3558691Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3558899Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3558985Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3559193Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3559281Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3559487Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3559576Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3559648Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3559706Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3559804Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3560125Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3560220Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3560431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3560519Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3560730Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3560815Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3560893Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3560949Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3561053Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3561388Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3561477Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3561690Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3561777Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3561986Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3562086Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3562164Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3562224Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3562325Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3562645Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3562721Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3562928Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3563017Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3563222Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3563314Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3563519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3563608Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3563816Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3563903Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3563980Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3564057Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3564159Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3564479Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3564559Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3564766Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3564856Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3565064Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3565155Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3565376Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3565480Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3565688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3565778Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3565853Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3565930Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3566028Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3566351Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3566428Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3566636Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3566724Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3566971Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3567061Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3567136Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3567197Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3567297Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3567620Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3567694Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3567905Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3567992Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3568223Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3568313Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3568388Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3568450Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3568548Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3568872Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3568947Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3569157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3569270Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3569503Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3569591Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3569802Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3569888Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3570113Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3570201Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3570280Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3570341Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3570444Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3570768Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3570843Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3571056Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3571144Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3571355Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3571444Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3571656Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3571743Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3571955Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3572043Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3572135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3572195Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3572299Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3572619Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3572696Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3572909Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3572995Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3573204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3573306Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3573525Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3573611Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3573821Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3573908Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3574000Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3574060Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3574163Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3574483Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3574562Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3574771Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3574860Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3575067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3575158Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3575236Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3575294Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3575395Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3575714Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3575791Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3576001Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3576107Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3576317Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3576409Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3576484Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3576547Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3576646Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3577006Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3577082Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3577311Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3577412Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3577620Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3577710Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3577931Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3578021Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3578229Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3578319Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3578396Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3578457Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3578557Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3578880Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3578955Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3579165Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3579254Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3579464Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3579550Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3579627Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3579686Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3579788Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3580109Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3589043Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3589256Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3589343Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3589557Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3589645Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3589724Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3589782Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3589885Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3597452Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3597550Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3597759Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3597849Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3598085Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3598177Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3598257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3598315Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3598417Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3598736Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3598815Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3599026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3599117Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3599326Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3599418Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3599501Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3599556Z Traceback (most recent call last): 2025-12-04T09:54:35.3599683Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3599731Z self.common( 2025-12-04T09:54:35.3599823Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3599875Z return func(*args, **kwds) 2025-12-04T09:54:35.3600029Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.3600073Z check_model( 2025-12-04T09:54:35.3600191Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3600236Z assert_equal_fn( 2025-12-04T09:54:35.3600377Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3600443Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3600604Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3600681Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3600736Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3600739Z 2025-12-04T09:54:35.3600792Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3600892Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3600990Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.3600992Z 2025-12-04T09:54:35.3601043Z The failure occurred for item [2] 2025-12-04T09:54:35.3601062Z 2025-12-04T09:54:35.3601146Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3601303Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3601306Z 2025-12-04T09:54:35.3601395Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3601472Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3601532Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3601647Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3601971Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3602050Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3602262Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3602355Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3602561Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3602652Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3602859Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3602950Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3603160Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3603248Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3603325Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3603382Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3603486Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3603807Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3603901Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3604111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3604203Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3604410Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3604502Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3604575Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3604638Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3604740Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3605076Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3605159Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3605371Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3605462Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3605682Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3605774Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3605982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3606073Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3606284Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3606376Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3606449Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3606509Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3606609Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3606965Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3607376Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3607893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3608060Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3608433Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3608594Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3612925Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3613030Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3613210Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3613758Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3613887Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3614255Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3614402Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3614753Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3614903Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3615072Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3615207Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3615377Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3615918Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3616072Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3616414Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3616565Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3617068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3617213Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3617556Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3617698Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3618046Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3618190Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3618317Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3618424Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3618594Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3619114Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3619243Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3619582Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3619771Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3620112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3620261Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3620602Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3620746Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3621084Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3621199Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3621291Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3621363Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3621513Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3621927Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3622023Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3622281Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3622408Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3622669Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3622783Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3623041Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3623150Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3623407Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3623514Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3623610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3623682Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3623815Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3624214Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3624308Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3624572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3624683Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3624942Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3625066Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3625325Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3625437Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3625695Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3625807Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3625891Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3625969Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3626088Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3626506Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3626614Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3626909Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3627017Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3628158Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3629155Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3629657Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3629870Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3630499Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3630702Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3630896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3631071Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3631300Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3632022Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3632196Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3632672Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3632872Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3633328Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3633522Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3633686Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3633883Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3634104Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3634816Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3634978Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3635435Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3635625Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3636079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3636271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3637052Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3637274Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3637761Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3637947Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3638106Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3638279Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3638498Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3639206Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3639370Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3639827Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3640015Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3640466Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3640653Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3641084Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3641208Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3641526Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3641654Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3641764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3641859Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3642012Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3642494Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3642672Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3642988Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3643123Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3643440Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3643564Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3643882Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3644012Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3644360Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3644510Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3644622Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3644710Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3644860Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3645353Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3645473Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3645793Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3645918Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3646231Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3646364Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3646474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3646568Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3646723Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3648220Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3648557Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3649206Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3649470Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3650065Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3650315Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3651023Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3651125Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3651352Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3651456Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3651547Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3651617Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3651742Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3652106Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3652237Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3652513Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3652612Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3652839Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3652937Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3653206Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3653307Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3653540Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3653639Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3653719Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3653791Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3653903Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3654269Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3654358Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3654592Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3654693Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3654927Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3655026Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3655104Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3655176Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3655301Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3655661Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3655746Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3655976Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3656073Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3656306Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3656404Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3656484Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3656556Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3656729Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3657184Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3657268Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3657518Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3657616Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3657851Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3657949Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3658179Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3658276Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3658509Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3658607Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3658688Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3658761Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3658874Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3659230Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3659314Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3659548Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3659646Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3659879Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3659997Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3660229Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3660326Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3660562Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3660659Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3660741Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3660811Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3660913Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3661246Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3661334Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3661542Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3661631Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3661849Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3661937Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3662013Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3662072Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3662175Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3662493Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3662569Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3662777Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3662866Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3663072Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3663163Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3663237Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3663300Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3663398Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3663722Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3663823Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3664033Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3664123Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3664332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3664421Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3664627Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3664716Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3664922Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3665013Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3665098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3665174Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3665274Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3665596Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3665680Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3665891Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3665978Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3666188Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3666277Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3666484Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3666573Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3666850Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3666940Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3667013Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3667076Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3667176Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3667495Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3667570Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3667781Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3667892Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3668104Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3668191Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3668401Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3668491Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3668699Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3668789Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3668865Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3668925Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3669037Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3669370Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3669446Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3669658Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3669756Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3669967Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3670054Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3670131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3670189Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3670290Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3670608Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3670686Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3670913Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3671001Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3671216Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3671303Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3671380Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3671439Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3671541Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3671864Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3671956Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3672165Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3672256Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3672463Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3672554Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3672763Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3672853Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3673073Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3673170Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3673249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3673307Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3673410Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3673741Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3673820Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3674029Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3674121Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3674328Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3674418Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3674492Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3674552Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3674652Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3674974Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3675049Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3675262Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3675352Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3675559Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3675651Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3675736Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3675796Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3675896Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3676220Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3676294Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3676507Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3676598Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3676852Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3676939Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3677032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3677101Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3677204Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3677522Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3677618Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3677836Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3677924Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3678137Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3678225Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3678303Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3678364Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3678468Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3678790Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3678870Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3679081Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3679174Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3679383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3679474Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3679684Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3679794Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3680006Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3680094Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3680180Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3680233Z Traceback (most recent call last): 2025-12-04T09:54:35.3680373Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3680415Z self.common( 2025-12-04T09:54:35.3680514Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3680561Z return func(*args, **kwds) 2025-12-04T09:54:35.3680695Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.3680735Z check_model( 2025-12-04T09:54:35.3680857Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3680898Z assert_equal_fn( 2025-12-04T09:54:35.3681060Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3681135Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3681301Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3681377Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3681437Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3681442Z 2025-12-04T09:54:35.3681490Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3681614Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3681723Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.3681726Z 2025-12-04T09:54:35.3681777Z The failure occurred for item [2] 2025-12-04T09:54:35.3681780Z 2025-12-04T09:54:35.3681857Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3682023Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3682025Z 2025-12-04T09:54:35.3682117Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3682199Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3682260Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3682365Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3682691Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3682773Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3682992Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3683083Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3683294Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3683382Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3683594Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3683694Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3683906Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3683994Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3684072Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3684131Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3684236Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3684560Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3684639Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3684859Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3684976Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3685187Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3685275Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3685352Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3685425Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3685528Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3685851Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3685930Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3686138Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3686228Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3686435Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3686525Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3686732Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3686937Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3687145Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3687236Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3687312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3687374Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3687478Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3687800Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3687902Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3688111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3688201Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3688408Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3688499Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3688574Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3688636Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3688736Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3689079Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3689167Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3689378Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3689466Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3689689Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3689782Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3689857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3689922Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3690022Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3690347Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3690423Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3690636Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3690726Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3690938Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3691027Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3691239Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3691327Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3691540Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3691640Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3691718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3691779Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3691883Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3692209Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3692284Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3692496Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3692583Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3692794Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3692890Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3693116Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3693205Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3693415Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3693512Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3693590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3693651Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3693755Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3694075Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3694155Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3694368Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3694457Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3694668Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3694756Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3694967Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3695056Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3695266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3695353Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3695430Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3695492Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3695609Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3695927Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3696006Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3696213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3696306Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3696516Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3696604Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3696867Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3696969Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3697179Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3697267Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3697344Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3697404Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3697519Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3697840Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3697921Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3698130Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3698221Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3698427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3698518Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3698727Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3698818Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3699029Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3699118Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3699196Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3699254Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3699356Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3699678Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3699772Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3699982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3700074Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3700281Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3700372Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3700448Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3700510Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3700611Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3700946Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3701034Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3701243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3701334Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3701550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3701642Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3701850Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3701941Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3702147Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3702238Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3702313Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3702377Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3702478Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3702804Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3702880Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3703092Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3703180Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3703395Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3703487Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3703703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3703793Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3704000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3704091Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3704166Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3704228Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3704328Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3704651Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3704727Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3704949Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3705047Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3705257Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3705344Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3705564Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3705656Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3705864Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3705956Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3706030Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3706091Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3706191Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3706514Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3706590Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3706839Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3706927Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3707136Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3707224Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3707301Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3707361Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3707464Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3707810Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3707889Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3708101Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3708188Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3708398Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3708486Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3708695Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3708799Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3709021Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3709108Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3709187Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3709248Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3709350Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3709684Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3709765Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3709974Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3710065Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3710276Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3710363Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3710573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3710664Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3710875Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3710963Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3711040Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3711099Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3711202Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3711524Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3711616Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3711825Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3711918Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3712124Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3712214Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3712288Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3712351Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3712450Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3712784Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3712871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3713079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3713170Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3713377Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3713680Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3713756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3713819Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3713921Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3714245Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3714320Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3714531Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3714620Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3714832Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3714922Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3715130Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3715221Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3715426Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3715517Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3715591Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3715669Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3715769Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3716093Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3716169Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3716380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3716468Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3716680Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3716800Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3717026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3717130Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3717341Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3717431Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3717509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3717583Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3717683Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3718009Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3718084Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3718297Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3718385Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3722002Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3722105Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3722185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3722247Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3722351Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3722681Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3722759Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3722975Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3723067Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3723308Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3723400Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3723474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3723539Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3723639Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3723965Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3724041Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3724253Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3724352Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3724573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3724661Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3724871Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3724963Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3725183Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3725277Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3725352Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3725416Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3725516Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3725840Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3725915Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3726127Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3726216Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3726426Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3726515Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3726726Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3726854Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3727066Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3727181Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3727259Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3727321Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3727423Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3727747Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3727822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3728036Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3728123Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3728334Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3728434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3728656Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3728743Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3728953Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3729053Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3729130Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3729189Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3729291Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3729614Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3729692Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3729903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3729990Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3730199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3730287Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3730364Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3730423Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3730524Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3730843Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3730919Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3731129Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3731233Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3731442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3731532Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3731606Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3731667Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3731767Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3732092Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3732169Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3732389Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3732489Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3732698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3732788Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3733006Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3733099Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3733306Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3733399Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3733474Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3733535Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3733635Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3733960Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3734039Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3734247Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3734339Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3734545Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3734635Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3734709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3734769Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3734870Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3735202Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3735278Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3735490Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3735577Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3735786Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3735874Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3735950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3736009Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3736112Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3736446Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3736531Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3736785Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3736873Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3737099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3737189Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3737267Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3737325Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3737427Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3737744Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3737821Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3738029Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3738120Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3738327Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3738418Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3738491Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3738554Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3738655Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3738975Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3739065Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3739274Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3739365Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3739572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3739661Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3739869Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3739959Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3740164Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3740267Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3740353Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3740413Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3740514Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3740854Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3740933Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3741144Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3741235Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3741446Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3741536Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3741620Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3741674Z Traceback (most recent call last): 2025-12-04T09:54:35.3741808Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3741853Z self.common( 2025-12-04T09:54:35.3741946Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3741996Z return func(*args, **kwds) 2025-12-04T09:54:35.3742126Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.3742168Z check_model( 2025-12-04T09:54:35.3742288Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3742331Z assert_equal_fn( 2025-12-04T09:54:35.3742473Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3742540Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3742702Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3742781Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3747200Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3747204Z 2025-12-04T09:54:35.3747254Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3747354Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3747454Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.3747457Z 2025-12-04T09:54:35.3747505Z The failure occurred for item [2] 2025-12-04T09:54:35.3747507Z 2025-12-04T09:54:35.3747586Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3747744Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3747746Z 2025-12-04T09:54:35.3747840Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3747918Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3747983Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3748087Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3748430Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3748525Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3748737Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3748830Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3749053Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3749147Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3749356Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3749447Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3749654Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3749745Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3749819Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3749881Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3749982Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3750307Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3750386Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3750595Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3750686Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3750894Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3750985Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3751073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3751136Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3751236Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3751561Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3751636Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3751847Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3751938Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3752151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3752239Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3752461Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3752560Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3752769Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3752859Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3752946Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3753008Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3753109Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3753433Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3753508Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3753719Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3753807Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3754017Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3754105Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3754182Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3754240Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3754343Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3754664Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3754741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3754953Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3755053Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3755264Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3755353Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3755429Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3755489Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3755591Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3755915Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3755993Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3756214Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3756315Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3756526Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3756616Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3756865Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3756972Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3757183Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3757271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3757348Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3757408Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3757511Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3757831Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3757910Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3758119Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3758210Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3758417Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3758508Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3758715Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3758806Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3759015Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3759117Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3759192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3759255Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3759355Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3759677Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3759754Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3759964Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3760055Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3760275Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3760383Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3760589Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3760679Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3760886Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3760986Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3761063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3761125Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3761228Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3761555Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3761632Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3761841Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3761931Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3762139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3762230Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3762437Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3762527Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3762735Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3762825Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3762901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3762974Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3763074Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3763395Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3763472Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3763685Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3763773Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3763985Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3764077Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3764295Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3764395Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3764603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3764693Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3764767Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3764840Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3764940Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3765263Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3765338Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3765550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3765637Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3765851Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3765939Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3766017Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3766076Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3766179Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3766503Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3766578Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3766828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3766915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3767140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3767228Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3767439Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3767526Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3767736Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3767823Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3767901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3767962Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3768064Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3768397Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3768487Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3768699Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3768787Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3769011Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3769100Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3769310Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3769398Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3769607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3769695Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3769771Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3769831Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3769934Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3770257Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3770335Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3770544Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3770634Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3770844Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3770944Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3771153Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3771241Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3771452Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3771543Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3771617Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3771675Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3771774Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3772095Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3772182Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3772398Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3772485Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3772693Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3772792Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3772866Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3772926Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3773024Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3773342Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3773418Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3773626Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3773713Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3773920Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3774007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3774213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3774300Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3774508Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3774593Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3774669Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3774728Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3774842Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3775160Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3775236Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3775442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3775529Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3775735Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3775823Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3776044Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3776140Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3776349Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3776435Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3776509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3776566Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3776674Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3777035Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3777111Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3777319Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3777406Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3777614Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3777701Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3777774Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3777832Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3777930Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3778250Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3778324Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3778532Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3778621Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3778827Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3778932Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3779005Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3779066Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3779164Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3779486Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3779560Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3779768Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3779856Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3780077Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3780176Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3780383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3780469Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3780688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3780777Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3780849Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3780908Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3781007Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3781326Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3781399Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3781609Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3781696Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3781903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3781990Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3782195Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3782281Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3782489Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3782576Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3782662Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3782719Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3782818Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3783141Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3783215Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3783424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3783512Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3783725Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3783813Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3783898Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3783964Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3784063Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3784380Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3784466Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3784675Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3784762Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3784969Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3785056Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3785129Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3785187Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3785285Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3785604Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3785680Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3785889Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3785977Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3786182Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3786269Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3786476Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3786577Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3786820Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3786908Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3786983Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3787041Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3787140Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3787458Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3787533Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3787755Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3787855Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3788064Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3788151Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3788356Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3788457Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3788663Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3788751Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3788824Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3788883Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3788981Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3789298Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3789374Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3789581Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3789667Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3789875Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3789962Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3790167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3790253Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3790459Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3790566Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3790639Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3790697Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3790796Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3791115Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3791188Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3791396Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3791484Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3791701Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3791798Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3791871Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3791928Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3792026Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3792356Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3792432Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3792644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3792731Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3792938Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3793024Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3793097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3793156Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3793256Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3793577Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3793651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3793861Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3793947Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3794154Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3794241Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3794459Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3794546Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3794755Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3794840Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3794914Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3794970Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3795069Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3795388Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3795473Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3795688Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3795775Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3795984Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3796070Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3796154Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3796212Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3796311Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3796628Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3796703Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3796949Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3797036Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3797245Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3797332Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3797403Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3797461Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3797560Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3797879Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3797952Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3798164Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3798271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3798478Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3798565Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3798638Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3798695Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3798793Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3799115Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3799189Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3799410Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3799508Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3799714Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3799800Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3799873Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3799932Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3800042Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3800364Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3800439Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3800647Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3800733Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3800938Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3801024Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3801234Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3801320Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3801529Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3801614Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3801687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3801743Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3801842Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3802162Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3802246Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3802455Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3802543Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3802750Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3802836Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3802909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3802969Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3803067Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3803397Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3803482Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3803689Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3803775Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3803998Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3804087Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3804295Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3804383Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3804588Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3804675Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3804759Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3804807Z Traceback (most recent call last): 2025-12-04T09:54:35.3804938Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3804977Z self.common( 2025-12-04T09:54:35.3805065Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3805110Z return func(*args, **kwds) 2025-12-04T09:54:35.3805240Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.3805277Z check_model( 2025-12-04T09:54:35.3805396Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3805434Z assert_equal_fn( 2025-12-04T09:54:35.3805574Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3805637Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3805800Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3805883Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3805938Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3805940Z 2025-12-04T09:54:35.3805985Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3806085Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3806180Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.3806182Z 2025-12-04T09:54:35.3806228Z The failure occurred for item [2] 2025-12-04T09:54:35.3806230Z 2025-12-04T09:54:35.3806304Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3806463Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3806466Z 2025-12-04T09:54:35.3806553Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3806630Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3806690Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3806824Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3807160Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3807249Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3807460Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3807548Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3807768Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3807855Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3808062Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3808149Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3808358Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3808443Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3808517Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3808575Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3808675Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3808995Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3809071Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3809279Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3809366Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3809572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3809673Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3809749Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3809807Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3809907Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3810228Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3810302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3810512Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3810600Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3810805Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3810902Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3811117Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3811204Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3811409Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3811505Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3811578Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3811637Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3811735Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3812055Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3812131Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3812339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3812426Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3812633Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3812721Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3812795Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3812853Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3812951Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3813269Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3813342Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3813552Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3813649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3813858Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3813944Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3814018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3814076Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3814174Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3814496Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3814570Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3814790Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3814886Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3815097Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3815183Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3815399Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3815487Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3815694Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3815781Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3815856Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3815914Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3816014Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3816333Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3816409Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3816617Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3816704Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3816953Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3817043Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3817260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3817350Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3817566Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3817673Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3817756Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3817820Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3817927Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3818248Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3818332Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3818543Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3818640Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3818873Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3818983Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3819200Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3819290Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3819523Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3819616Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3819699Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3819763Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3819871Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3820194Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3820277Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3820489Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3820585Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3820796Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3820892Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3821107Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3821197Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3821414Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3821505Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3821600Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3821663Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3821772Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3822097Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3822181Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3822392Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3822488Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3822701Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3822797Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3823020Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3823127Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3823338Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3823432Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3823525Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3823588Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3823698Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3824024Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3824110Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3824323Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3824420Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3824630Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3824728Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3824805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3824875Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3824977Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3825304Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3825382Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3825600Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3825705Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3825916Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3826013Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3826224Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3826320Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3826532Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3826629Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3826707Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3826806Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3826923Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3827263Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3827340Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3827556Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3827658Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3827871Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3827959Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3828173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3828265Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3828473Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3828565Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3828643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3828711Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3828814Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3829145Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3829224Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3829442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3829533Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3829749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3829852Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3830068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3830160Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3830375Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3830471Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3830548Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3830616Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3830719Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3831062Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3831150Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3831369Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3831459Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3831674Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3831775Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3831857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3831918Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3832024Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3832347Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3832426Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3832638Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3832731Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3832947Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3833036Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3833254Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3833342Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3833556Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3833645Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3833726Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3833800Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3833904Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3834224Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3834304Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3834513Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3834606Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3834815Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3834907Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3835132Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3835229Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3835441Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3835529Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3835608Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3835676Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3835781Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3836102Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3836180Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3836388Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3836479Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3836686Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3836814Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3836889Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3836951Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3837051Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3837374Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3837452Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3837662Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3837753Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3837974Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3838066Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3838141Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3838204Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3838304Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3838629Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3838705Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3838917Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3839017Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3839242Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3839330Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3839540Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3839631Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3839853Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3839946Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3840022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3840087Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3840187Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3840510Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3840585Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3840798Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3840887Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3841099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3841187Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3841398Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3841485Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3841697Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3841797Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3841873Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3841934Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3842035Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3842360Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3842434Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3842646Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3842735Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3842944Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3843042Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3843128Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3843186Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3843287Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3843615Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3843693Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3843903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3843993Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3844204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3844290Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3844368Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3844427Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3844530Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3844849Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3844929Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3845136Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3845223Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3845428Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3845515Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3845721Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3845821Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3846027Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3846114Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3846186Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3846246Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3846344Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3846664Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3846773Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3846993Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3847091Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3847297Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3847383Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3847609Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3847697Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3847903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3847990Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3848063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3848122Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3848221Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3848541Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3848616Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3848824Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3848912Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3849119Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3849206Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3849411Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3849498Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3849703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3849810Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3849883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3849941Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3850039Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3850359Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3850432Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3850640Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3850725Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3850945Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3851042Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3851116Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3851174Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3851271Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3851600Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3851675Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3851885Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3851971Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3852179Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3852265Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3852339Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3852399Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3852499Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3852821Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3852897Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3853107Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3853193Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3853403Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3853501Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3853711Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3853798Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3854006Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3854091Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3854165Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3854221Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3854322Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3854643Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3857678Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3857902Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3857989Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3858195Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3858296Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3858371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3858428Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3858528Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3858848Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3858922Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3859129Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3859216Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3859421Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3859509Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3859582Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3859641Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3859737Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3860054Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3860127Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3860335Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3860438Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3860648Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3860736Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3860809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3860867Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3860964Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3861284Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3861358Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3861582Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3861679Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3861886Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3861972Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3862045Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3862117Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3862217Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3862535Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3862609Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3862817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3862903Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3863110Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3863196Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3863402Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3863488Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3863695Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3863780Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3863854Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3863910Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3864010Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3864339Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3864420Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3864628Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3864714Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3864919Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3865007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3865080Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3865139Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3865237Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3865566Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3865649Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3865855Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3865961Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3866167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3866255Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3866462Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3866550Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3866791Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3866878Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3866950Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3867010Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3867114Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3867441Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3867520Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3867728Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3867819Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3868028Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3868136Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3868344Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3868435Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3868642Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3868731Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3868805Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3868867Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3868968Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3869307Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3869393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3869605Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3869695Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3869903Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3870006Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3870218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3870308Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3870518Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3870609Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3870693Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3870746Z Traceback (most recent call last): 2025-12-04T09:54:35.3870875Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3870917Z self.common( 2025-12-04T09:54:35.3871009Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3871057Z return func(*args, **kwds) 2025-12-04T09:54:35.3871185Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.3871227Z check_model( 2025-12-04T09:54:35.3871346Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3871390Z assert_equal_fn( 2025-12-04T09:54:35.3871529Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3871594Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3871754Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3871831Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3871887Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3871890Z 2025-12-04T09:54:35.3871951Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3872057Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3872166Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.3872169Z 2025-12-04T09:54:35.3872217Z The failure occurred for item [2] 2025-12-04T09:54:35.3872222Z 2025-12-04T09:54:35.3872295Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3872452Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3872455Z 2025-12-04T09:54:35.3872543Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3872621Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3872682Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3872785Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3873119Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3873208Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3873419Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3873512Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3873731Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3873824Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3874033Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3874124Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3874332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3874424Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3874502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3874561Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3874665Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3874985Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3875065Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3875275Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3875366Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3875574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3875664Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3875738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3875813Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3875913Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3876237Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3876313Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3876527Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3876615Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3876862Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3876953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3877173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3877278Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3877486Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3877577Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3877651Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3877725Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3877826Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3878148Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3878224Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3878435Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3878522Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3878733Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3878821Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3878899Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3878957Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3879060Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3879383Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3879457Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3879669Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3879756Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3879980Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3880067Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3880144Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3880205Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3880306Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3880628Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3880706Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3880916Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3881023Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3881249Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3881340Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3881552Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3881640Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3881863Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3881952Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3882030Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3882090Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3882193Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3882512Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3882589Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3882799Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3882891Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3883099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3883190Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3883399Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3883489Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3883700Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3883800Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3883877Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3883937Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3884040Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3884361Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3884440Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3884650Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3884742Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3884948Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3885049Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3885265Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3885356Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3885565Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3885676Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3885753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3885815Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3885915Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3886238Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3886316Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3886525Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3886615Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3886859Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3886950Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3887157Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3887249Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3887455Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3887544Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3887619Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3887681Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3887799Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3888125Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3888204Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3888414Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3888504Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3888711Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3888802Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3889021Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3889123Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3889330Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3889420Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3889495Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3889555Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3889666Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3889997Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3890077Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3890296Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3890386Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3890602Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3890699Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3890776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3890844Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3890947Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3891278Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3891356Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3891574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3891665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3891895Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3891985Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3892202Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3892293Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3892509Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3892599Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3892682Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3892745Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3892853Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3893191Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3893277Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3893495Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3893584Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3893811Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3893905Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3894123Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3894215Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3894431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3894522Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3894605Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3894667Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3894778Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3895107Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3895191Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3895409Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3895500Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3895718Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3895810Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3896037Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3896127Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3896343Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3896433Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3896515Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3896577Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3896684Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3897041Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3897147Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3897374Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3897470Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3897685Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3897775Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3897870Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3897933Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3898041Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3898364Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3898448Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3898657Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3898753Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3898965Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3899061Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3899271Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3899367Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3899580Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3899675Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3899751Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3899819Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3899926Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3900271Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3900354Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3900564Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3900659Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3900867Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3900961Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3901170Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3901278Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3901497Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3901593Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3901672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3901739Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3901842Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3902184Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3902269Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3902480Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3902577Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3902788Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3902884Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3902964Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3903031Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3903134Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3903463Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3903541Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3903758Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3903849Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3904067Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3904171Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3904254Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3904318Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3904425Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3904752Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3904830Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3905047Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3905138Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3905363Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3905463Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3905678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3905768Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3905995Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3906086Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3906170Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3906232Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3906341Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3906667Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3906784Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3907003Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3907094Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3907310Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3907401Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3907617Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3907707Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3907922Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3908012Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3908095Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3908186Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3908293Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3908616Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3908701Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3908915Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3909010Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3909221Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3909318Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3909416Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3909490Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3909599Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3909921Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3910004Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3910227Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3910325Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3910535Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3910632Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3910709Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3910778Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3910879Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3911211Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3911290Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3911508Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3911604Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3911814Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3911909Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3912120Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3912216Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3912453Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3912549Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3912626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3912693Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3912795Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3913123Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3913200Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3913418Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3913519Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3913747Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3913838Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3914054Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3914151Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3914369Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3914458Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3914531Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3914591Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3914688Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3915009Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3915083Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3917661Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3917761Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3917972Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3918063Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3918271Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3918357Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3918566Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3918682Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3918759Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3918818Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3918919Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3919241Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3919316Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3919524Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3919613Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3919819Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3919920Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3920005Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3920063Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3920161Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3920500Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3920578Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3920785Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3920874Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3921080Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3921167Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3921239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3921300Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3921399Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3921720Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3921794Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3922002Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3922088Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3922293Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3922380Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3922588Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3922686Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3922893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3922980Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3923053Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3923111Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3923209Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3923531Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3923605Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3923825Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3923920Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3924127Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3924212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3924287Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3924354Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3924454Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3924774Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3924849Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3925058Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3925143Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3925352Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3925438Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3925512Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3925567Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3925667Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3925985Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3926059Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3926266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3926366Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3926574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3926662Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3926737Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3926838Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3926937Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3927260Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3927335Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3927546Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3927648Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3927868Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3927954Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3928026Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3928086Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3928204Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3928523Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3928599Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3928812Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3928898Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3929111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3929197Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3929404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3929492Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3929698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3929785Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3929857Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3929914Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3930012Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3930331Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3930417Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3930629Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3930715Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3930923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3931009Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3931084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3931143Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3931244Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3931573Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3931656Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3931864Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3931950Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3932166Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3932254Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3932463Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3932551Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3932761Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3932847Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3932922Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3932981Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3933082Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3933405Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3933482Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3933694Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3933782Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3933990Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3934076Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3934285Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3934383Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3934592Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3934678Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3934753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3934811Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3934910Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3935228Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3935303Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3935522Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3935621Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3935829Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3935915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3936130Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3936220Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3936429Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3936516Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3936590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3936646Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3936776Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3937096Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3937170Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3937379Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3937467Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3937674Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3937762Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3937844Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.3937894Z Traceback (most recent call last): 2025-12-04T09:54:35.3938026Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.3938089Z self.common( 2025-12-04T09:54:35.3938178Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.3938224Z return func(*args, **kwds) 2025-12-04T09:54:35.3938352Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.3938391Z check_model( 2025-12-04T09:54:35.3938506Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.3938546Z assert_equal_fn( 2025-12-04T09:54:35.3938685Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.3938748Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.3938909Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.3938985Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.3939039Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.3939044Z 2025-12-04T09:54:35.3939089Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.3939188Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.3939294Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.3939308Z 2025-12-04T09:54:35.3939355Z The failure occurred for item [2] 2025-12-04T09:54:35.3939357Z 2025-12-04T09:54:35.3939431Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.3939588Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.3939590Z 2025-12-04T09:54:35.3939680Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.3939767Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3939828Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3939928Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3940247Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3940323Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3940536Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3940626Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3940833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3940922Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3941129Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3941216Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3941425Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3941511Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3941587Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3941646Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3941747Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3942080Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3942156Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3942364Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3942451Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3942658Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3942747Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3942821Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3942881Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3942990Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3943330Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3943405Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3943613Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3943714Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3943921Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3944009Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3944215Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3944301Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3944506Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3944592Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3944666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3944725Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3944823Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3945145Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3945219Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3945427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3945512Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3945721Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3945819Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3945892Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3945950Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3946048Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3946366Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3946439Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3946651Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3946737Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3947003Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3947101Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3947175Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3947233Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3947333Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3947674Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3947750Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3947959Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3948046Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3948252Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3948339Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3948546Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3948631Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3948838Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3948924Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3948998Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3949057Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3949156Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3949478Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3949553Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3949782Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3949871Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3950079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3950166Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3950373Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3950461Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3950669Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3950756Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3950842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3950913Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3951013Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3951331Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3951406Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3951629Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3951718Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3951925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3952012Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3952217Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3952303Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3952511Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3952598Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3952672Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3952730Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3952831Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3953148Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3953223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3953429Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3953528Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3953735Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3953822Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3954029Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3954116Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3954322Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3954409Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3954482Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3954542Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3954641Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3954969Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3955055Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3955262Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3955358Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3955563Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3955651Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3955857Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3955944Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3956151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3956237Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3956310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3956369Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3956466Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3956820Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3956895Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3957103Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3957191Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3957397Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3957506Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3957579Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3957638Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3957738Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3958061Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3958134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3958346Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3958433Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3958642Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3958747Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3958966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3959053Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3959257Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3959357Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3959431Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3959490Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3959589Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3959907Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3959981Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3960189Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3960276Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3960484Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3960572Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3960781Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3960868Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3961073Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3961159Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3961235Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3961292Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3961402Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3961720Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3961794Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3962003Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3962090Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3962296Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3962384Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3962603Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3962699Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3962908Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3962995Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3963069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3963127Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3963234Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3963555Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3963629Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3963837Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3963922Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3964129Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3964217Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3964290Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3964349Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3964447Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3964768Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3964841Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3965049Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3965135Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3965352Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3965438Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3965647Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3965733Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3965939Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3966025Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3966099Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3966157Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3966256Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3966586Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3966669Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3966904Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3966990Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3967219Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3967307Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3967516Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3967602Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3967810Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3967895Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3967969Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3968025Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3968126Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3968447Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3968522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3968731Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3968816Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3969024Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3969110Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3969198Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3969255Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3969356Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3969677Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3969751Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3969957Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3970045Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3970251Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3970351Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3970434Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3970494Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3970592Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3970913Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3970998Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3971206Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3971293Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3971500Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3971588Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3971794Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3971881Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3972087Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3972174Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3972245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3972306Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3972405Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3972723Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3972795Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3973004Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3973110Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3973317Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3973404Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3973610Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3973697Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3973904Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3973992Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3974065Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3974122Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3974232Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3974559Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3974632Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3974840Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3974937Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3975144Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3975231Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3975304Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3975361Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3975457Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3975775Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3975849Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3976058Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3976144Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3976350Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3976439Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3976513Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3976571Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3976670Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3977021Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3977110Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3977318Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3977407Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3977615Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3977701Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3977909Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3977995Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3978218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3978316Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3978390Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3978447Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3978546Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3978879Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3978954Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3979161Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3979248Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3979453Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3979539Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3979747Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3979832Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3980040Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3980127Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3980202Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3980260Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3980359Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3980680Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3980756Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3980976Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3981064Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3981270Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3981358Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3981564Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3981649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3981856Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3981943Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3982029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3982095Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3982194Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3982516Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3982590Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3982809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3982897Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3983105Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3983193Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3983265Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3983323Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3983421Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3983741Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3983815Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3984024Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3984113Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3984319Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3984406Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3984479Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3984538Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3984637Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3984970Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3985044Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3985253Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3985341Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3985552Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3985638Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3985848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3985947Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3986165Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3986251Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3986328Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3986387Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3986502Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3986854Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3986932Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3987145Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3987232Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3987441Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3987528Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3987603Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3987661Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3987764Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3988083Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3988159Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3988369Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3988457Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3988666Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3988771Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3988846Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3988903Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3989003Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3989323Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3989398Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3989606Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3989694Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3989916Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3990017Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3990090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3990150Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3990249Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3990581Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3990657Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3990872Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3990961Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3991167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3991255Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3991328Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3991389Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3991488Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3991812Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3991887Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3992097Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3992183Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3992392Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3992477Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3992698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3992784Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3992995Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3993082Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3993155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3993211Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.3993310Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.3993635Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.3993730Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3993952Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3994039Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3994249Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3994335Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3994421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3994482Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3994582Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3994903Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3994979Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3995189Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3995275Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3995484Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3995570Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3995783Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3995869Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3996080Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3996166Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3996239Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3996300Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3996414Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3996735Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3996853Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3997061Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3997151Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3997360Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3997449Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3997661Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3997770Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3997991Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3998077Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3998151Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.3998208Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.3998319Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.3998641Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.3998716Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.3998924Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3999012Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3999218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3999306Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3999511Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.3999599Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3999805Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.3999893Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.3999967Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4000023Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4000126Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4000453Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4000545Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4000755Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4000845Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4001057Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4001147Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4001221Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4001283Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4001383Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4001724Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4001816Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4002031Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4002121Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4002343Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4002433Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4002646Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4002737Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4002952Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4003041Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4003126Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.4003179Z Traceback (most recent call last): 2025-12-04T09:54:35.4003310Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.4003350Z self.common( 2025-12-04T09:54:35.4003440Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.4003487Z return func(*args, **kwds) 2025-12-04T09:54:35.4003618Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.4003660Z check_model( 2025-12-04T09:54:35.4003780Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.4003820Z assert_equal_fn( 2025-12-04T09:54:35.4003961Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.4004024Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.4004187Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.4004263Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.4004331Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4004333Z 2025-12-04T09:54:35.4004380Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4004479Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4004577Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4004580Z 2025-12-04T09:54:35.4004627Z The failure occurred for item [2] 2025-12-04T09:54:35.4004629Z 2025-12-04T09:54:35.4004706Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4004864Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4004868Z 2025-12-04T09:54:35.4004958Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4005037Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4005100Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4005207Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4005547Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4005636Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4005850Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4005943Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4006166Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4006261Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4006475Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4006566Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4006819Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4006911Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4006986Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4007047Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4007149Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4007482Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4007563Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4007778Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4007871Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4008082Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4008173Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4008266Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4008328Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4008429Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4008761Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4008838Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4009055Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4009147Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4009365Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4009469Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4009696Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4009786Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4010000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4010090Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4010179Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4010242Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4010343Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4010674Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4010751Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4010966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4011055Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4011273Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4011362Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4011440Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4011500Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4011606Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4011934Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4012013Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4012232Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4012334Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4012550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4012642Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4012721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4012782Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4012885Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4013212Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4013292Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4013517Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4013620Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4013835Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4013927Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4014139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4014242Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4014458Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4014548Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4014628Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4014688Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4014789Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4015115Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4015197Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4015411Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4015503Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4015715Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4015806Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4016017Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4016107Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4016319Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4016425Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4016504Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4016564Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4016667Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4017057Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4017134Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4017348Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4017440Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4017671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4017774Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4017987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4018079Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4018305Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4018395Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4018471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4018531Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4018633Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4018965Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4019042Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4019254Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4019345Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4019560Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4019650Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4019863Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4019954Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4020167Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4020256Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4020331Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4020406Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4020507Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4020834Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4020910Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4021124Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4021213Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4021424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4021515Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4021738Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4021839Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4022050Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4022142Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4022218Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4022293Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4022393Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4022715Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4022791Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4023002Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4023090Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4023301Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4023390Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4023464Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4023524Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4023624Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4023950Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4024023Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4024235Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4024707Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4024917Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4025005Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4025216Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4025304Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4025515Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4025602Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4025679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4025739Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4025839Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4031192Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4031296Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4031513Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4031619Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4031836Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4031926Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4032145Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4032236Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4032449Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4032537Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4032613Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4032674Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4032777Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4033109Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4033189Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4033407Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4033497Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4033713Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4033820Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4034035Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4034125Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4034339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4034429Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4034505Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4034564Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4034667Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4035000Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4035093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4035320Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4035411Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4035624Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4035727Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4035806Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4035866Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4035969Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4036301Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4036380Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4036594Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4036684Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4036919Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4037010Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4037225Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4037316Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4037529Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4037619Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4037694Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4037756Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4037884Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4038219Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4038298Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4038512Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4038602Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4038816Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4038907Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4039135Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4039237Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4039451Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4039542Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4039618Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4039678Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4039794Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4040130Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4040207Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4040426Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4040518Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4040733Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4040825Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4040904Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4040968Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4041075Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4041427Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4041507Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4041737Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4041831Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4042081Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4042177Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4042261Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4042325Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4042431Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4042783Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4042866Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4043095Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4043190Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4043429Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4043533Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4043759Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4043853Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4044091Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4044187Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4044268Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4044332Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4044441Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4044790Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4044871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4045101Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4045197Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4045428Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4045523Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4045749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4045843Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4046069Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4046164Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4046259Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4046322Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4046430Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4046824Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4046907Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4047135Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4047232Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4047460Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4047580Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4047679Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4047743Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4047850Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4048199Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4048297Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4048527Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4048622Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4048848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4048944Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4049023Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4049087Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4049194Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4049546Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4049627Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4049856Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4049952Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4050181Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4050276Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4050502Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4050612Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4050839Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4050935Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4051016Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4051081Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4051190Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4051546Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4051628Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4051873Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4051979Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4052211Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4052306Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4052538Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4052652Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4052884Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4052981Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4053062Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4053127Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4053235Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4053592Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4053674Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4053906Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4054002Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4054232Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4054327Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4054556Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4054650Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4054881Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4054991Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4055074Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4055137Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4055247Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4055602Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4055684Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4055919Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4056016Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4056256Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4056361Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4056442Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4056505Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4056615Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4057029Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4057114Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4057348Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4057446Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4057677Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4057773Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4057855Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4057920Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4058030Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4058385Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4058469Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4058701Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4058797Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4059027Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4059124Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4059371Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4059470Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4059701Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4059797Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4059878Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4059943Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4060051Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4060409Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4060515Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4060762Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4060858Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4061087Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4061183Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4061276Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4061341Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4061448Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4061805Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4061888Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4062119Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4062214Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4062447Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4062542Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4062624Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4062690Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4062799Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4063150Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4063230Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4063462Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4063570Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4063802Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4063897Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4063978Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4064040Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4064149Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4064503Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4064589Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4064836Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4064943Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4065171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4065265Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4065345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4065411Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4065533Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4065888Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4065970Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4066198Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4066293Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4066519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4066615Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4066883Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4066981Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4067209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4067305Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4067384Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4067448Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4067555Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4067907Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4068014Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4068245Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4068343Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4068572Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4068667Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4068749Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4068816Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4068923Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4069303Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4069413Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4069646Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4069742Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4069989Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4070087Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4070315Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4070412Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4070641Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4070736Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4070816Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4070882Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4070990Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4071346Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4071427Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4071657Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4071752Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4071982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4072078Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4072322Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4072417Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4072647Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4072742Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4072822Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4072886Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4072995Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4073349Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4073443Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4073683Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4073777Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4074003Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4074098Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4074338Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4074434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4074665Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4074760Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4074841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4074903Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4075011Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4075363Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4075444Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4075673Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4075768Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4075994Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4076088Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4076169Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4076233Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4076355Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4076706Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4076839Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4077068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4077164Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4077389Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4077485Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4077711Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4077828Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4078069Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4078163Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4078243Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4078307Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4078441Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4078793Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4078875Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4079104Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4079200Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4079428Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4079525Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4079751Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4079846Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4080073Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4080168Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4080249Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4080312Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4080420Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4080771Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4080867Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4081095Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4081191Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4081417Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4081512Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4081741Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4081839Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4082081Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4082187Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4082276Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.4082331Z Traceback (most recent call last): 2025-12-04T09:54:35.4082472Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.4082514Z self.common( 2025-12-04T09:54:35.4082610Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.4082660Z return func(*args, **kwds) 2025-12-04T09:54:35.4082810Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.4082853Z check_model( 2025-12-04T09:54:35.4082981Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.4083023Z assert_equal_fn( 2025-12-04T09:54:35.4083176Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.4083243Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.4083419Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.4083499Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.4083558Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4083560Z 2025-12-04T09:54:35.4083609Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4083717Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4083821Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4083823Z 2025-12-04T09:54:35.4083873Z The failure occurred for item [2] 2025-12-04T09:54:35.4083875Z 2025-12-04T09:54:35.4083956Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4084127Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4084130Z 2025-12-04T09:54:35.4084226Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4084308Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4084371Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4084482Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4087748Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4087870Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4088101Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4088201Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4088428Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4088522Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4088750Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4088844Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4089101Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4089212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4089293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4089356Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4089464Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4089835Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4089921Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4090155Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4090253Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4090482Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4090579Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4090658Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4090725Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4090835Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4091189Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4091272Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4091502Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4091597Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4091824Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4091920Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4092160Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4092257Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4092484Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4092579Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4092658Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4092722Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4092828Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4093179Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4093275Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4093514Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4093609Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4093834Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4093929Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4094022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4094086Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4094193Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4094542Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4094624Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4094854Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4094947Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4095175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4095269Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4095349Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4095415Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4095523Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4095873Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4095954Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4096184Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4096294Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4096522Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4096617Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4096870Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4096965Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4097196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4097290Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4097373Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4097438Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4097565Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4097931Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4098013Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4098239Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4098348Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4098576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4098671Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4098900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4098993Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4099219Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4099313Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4099395Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4099459Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4099568Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4099918Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4100001Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4100228Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4100323Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4100549Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4100659Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4100887Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4100981Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4101208Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4101302Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4101383Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4101447Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4101557Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4101923Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4102018Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4102246Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4102342Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4102583Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4102678Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4102906Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4103002Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4103232Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4103325Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4103406Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4103469Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4103579Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4103930Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4104014Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4104243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4104341Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4104571Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4104667Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4104894Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4105009Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4105236Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4105332Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4105411Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4105475Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4105583Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4105934Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4106016Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4106256Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4106365Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4106591Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4106686Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4106796Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4106878Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4106988Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4107339Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4107420Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4107647Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4107741Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4107969Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4108064Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4108292Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4108388Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4108614Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4108709Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4108789Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4108853Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4108961Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4109333Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4109415Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4109644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4109739Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4109966Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4110061Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4110289Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4110385Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4110627Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4110737Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4110816Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4110882Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4110989Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4111355Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4111437Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4111669Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4111766Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4111995Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4112089Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4112318Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4112413Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4112643Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4112738Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4112819Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4112883Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4112990Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4113343Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4113436Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4113668Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4113764Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4113996Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4114092Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4114174Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4114239Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4114349Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4114702Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4114796Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4115037Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4115134Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4115362Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4115469Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4115698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4115795Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4116026Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4116122Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4116204Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4116267Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4116378Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4116734Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4119297Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4119538Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4119637Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4119867Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4119964Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4120194Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4120323Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4120553Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4120651Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4120733Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4120797Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4120909Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4121270Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4121355Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4121604Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4121715Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4121944Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4122037Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4122118Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4122180Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4122312Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4122669Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4122750Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4122981Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4123075Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4123304Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4123399Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4123484Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4123548Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4123659Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4124014Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4124096Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4124325Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4124422Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4124663Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4124759Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4124987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4125082Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4125309Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4125404Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4125486Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4125551Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4125661Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4126025Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4126120Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4126351Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4126447Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4126686Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4126816Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4127044Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4127141Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4127366Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4127462Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4127543Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4127606Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4127716Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4128070Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4128152Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4128381Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4128476Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4128703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4128800Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4128896Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4128959Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4129068Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4129421Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4129502Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4129731Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4129827Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4130056Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4130168Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4130262Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4130328Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4130435Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4130792Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4130887Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4131121Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4131218Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4131447Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4131541Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4131772Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4131867Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4132097Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4132193Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4132274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4132340Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4132449Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4132801Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4132882Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4133112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4133221Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4133451Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4133547Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4133777Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4133871Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4134100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4134195Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4134278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4134342Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4134463Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4134825Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4134907Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4135137Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4135244Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4135475Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4135570Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4135799Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4135893Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4136122Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4136216Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4136298Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4136361Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4136470Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4136856Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4136939Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4137166Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4137263Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4137491Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4137606Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4137688Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4137751Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4137859Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4138212Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4138297Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4138530Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4138627Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4138871Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4138987Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4139066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4139132Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4139239Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4139606Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4139691Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4139923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4140021Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4140252Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4140349Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4140578Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4140675Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4140905Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4141002Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4141084Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4141148Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4141256Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4141611Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4141693Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4141935Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4142031Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4142260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4142356Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4142436Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4142499Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4142605Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4142958Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4143050Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4143296Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4143391Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4143619Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4143714Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4143808Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4143871Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4143980Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4144333Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4144415Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4144644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4144738Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4144967Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4145063Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4145144Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4145207Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4145316Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4145670Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4145752Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4145983Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4146092Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4146321Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4146418Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4146499Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4146565Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4146672Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4147066Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4147148Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4147396Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4147505Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4147733Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4147829Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4148072Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4148169Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4148400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4148496Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4148577Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4148640Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4148748Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4149101Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4149183Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4149413Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4149510Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4149739Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4149834Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4149915Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4149980Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4150088Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4150443Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4150540Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4150773Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4150867Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4151097Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4151192Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4151422Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4151519Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4151759Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4151867Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4151949Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4152014Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4152125Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4152493Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4152576Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4152809Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4152906Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4153137Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4153232Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4153464Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4153558Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4153791Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4153887Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4153970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4154034Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4154145Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4154500Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4154583Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4154830Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4154928Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4155160Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4155255Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4155487Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4155585Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4155819Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4155916Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4156019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4156093Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4156203Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4156557Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4156639Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4156923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4157022Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4157256Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4157352Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4157433Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4157498Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4157609Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4157969Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4158054Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4158289Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4158387Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4158617Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4158713Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4158946Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4159062Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4159291Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4159389Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4159471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4159537Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4159648Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4160004Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4160087Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4160319Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4160429Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4160674Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4160771Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4161000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4161116Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4161346Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4161444Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4161526Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4161592Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4161701Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4162058Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4162142Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4162373Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4162470Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4162701Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4162799Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4163031Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4163127Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4163358Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4163468Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4163550Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4163617Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4163728Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4164086Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4164168Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4164404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4164504Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4164748Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4164858Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4165089Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4165185Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4165417Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4165526Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4165619Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.4165677Z Traceback (most recent call last): 2025-12-04T09:54:35.4165823Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.4165867Z self.common( 2025-12-04T09:54:35.4165968Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.4166018Z return func(*args, **kwds) 2025-12-04T09:54:35.4166161Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.4166203Z check_model( 2025-12-04T09:54:35.4166333Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.4166377Z assert_equal_fn( 2025-12-04T09:54:35.4166534Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.4166604Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.4166822Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.4166906Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.4166968Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4166971Z 2025-12-04T09:54:35.4167024Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4167141Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4167257Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4167260Z 2025-12-04T09:54:35.4167311Z The failure occurred for item [2] 2025-12-04T09:54:35.4167313Z 2025-12-04T09:54:35.4167396Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4167570Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4167591Z 2025-12-04T09:54:35.4167690Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4167776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4167843Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4167955Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4168314Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4168397Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4168636Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4168735Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4168985Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4169096Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4169328Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4169425Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4169670Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4169766Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4169850Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4169914Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4170027Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4170387Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4170470Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4170706Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4170802Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4171038Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4171134Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4171217Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4171282Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4171392Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4171748Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4171832Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4172083Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4172181Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4172411Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4172507Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4172737Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4172834Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4173064Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4173162Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4173257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4173331Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4173443Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4173796Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4173878Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4174120Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4174218Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4174449Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4174546Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4174626Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4174690Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4174798Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4175152Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4175234Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4175466Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4175564Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4175795Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4175891Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4175971Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4176038Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4176145Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4176515Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4176597Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4176871Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4176967Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4177197Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4177291Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4177520Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4177631Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4177874Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4177969Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4178052Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4178118Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4178241Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4178596Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4178680Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4178913Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4179010Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4179240Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4179335Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4179566Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4179662Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4179893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4179989Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4180071Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4180135Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4180244Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4180599Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4180698Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4180931Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4181027Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4181258Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4181353Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4181584Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4181679Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4181909Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4182016Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4182110Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4182174Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4182284Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4182648Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4182733Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4182963Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4183061Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4183292Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4183387Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4183616Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4183711Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4183940Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4184035Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4184117Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4184181Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4184290Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4184643Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4184727Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4184957Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4185065Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4185293Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4185390Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4185622Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4185716Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4185945Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4186041Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4186122Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4186195Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4186316Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4186668Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4186799Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4187055Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4187154Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4187383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4187480Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4187560Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4187626Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4187733Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4188093Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4188176Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4188407Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4188503Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4188730Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4188826Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4189053Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4189153Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4189406Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4189504Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4189585Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4189653Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4189761Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4190115Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4190197Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4190430Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4190526Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4190768Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4190878Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4191106Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4191201Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4191443Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4191541Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4191623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4191690Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4191800Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4192158Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4192239Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4192472Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4192568Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4192800Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4192898Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4193129Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4193225Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4193454Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4193551Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4193646Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4193710Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4193819Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4194175Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4194256Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4194488Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4194584Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4194817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4194924Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4195018Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4195082Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4195191Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4195548Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4195641Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4195877Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4195973Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4196202Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4196296Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4196524Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4196619Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4196897Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4196993Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4197075Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4197140Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4197250Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4197605Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4197687Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4197919Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4198034Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4198264Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4198362Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4198592Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4198687Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4198918Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4199014Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4199098Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4199161Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4199290Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4199659Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4199741Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4199977Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4200088Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4200320Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4200418Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4200502Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4200565Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4200678Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4201029Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4201113Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4201343Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4201441Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4201671Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4201768Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4201848Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4201914Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4202022Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4202380Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4202476Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4202711Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4202807Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4203037Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4203133Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4203363Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4203460Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4203703Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4203820Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4203901Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4203965Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4204074Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4204438Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4204520Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4204752Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4204848Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4205077Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4205172Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4205404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4205500Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4205728Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4205825Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4205907Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4205970Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4206079Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4206433Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4206516Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4206789Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4206885Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4207115Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4207210Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4207292Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4207355Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4207464Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4207818Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4207918Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4208163Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4208258Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4208487Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4208582Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4208677Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4208744Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4208853Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4209208Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4209291Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4209519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4209616Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4209844Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4209941Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4210172Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4210270Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4210501Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4210596Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4210676Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4210741Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4210850Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4211219Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4211303Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4211532Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4211629Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4211858Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4211953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4212182Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4212290Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4212529Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4212625Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4212706Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4212771Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4212889Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4213244Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4213328Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4213559Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4213655Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4213886Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4213982Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4214213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4214310Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4214541Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4214638Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4214720Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4214784Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4214893Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4215252Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4217237Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4217468Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4217565Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4217794Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4217890Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4217970Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4218035Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4218144Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4219166Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4225623Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4225855Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4225950Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4226201Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4226300Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4226382Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4226447Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4226557Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4226935Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4227017Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4227248Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4227343Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4227573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4227669Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4227902Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4227997Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4228224Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4228319Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4228400Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4228485Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4228594Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4228951Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4229035Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4229268Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4229363Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4229596Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4229693Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4229804Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4229882Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4229991Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4230342Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4230424Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4230675Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4230776Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4231004Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4231103Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4231189Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4231254Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4231364Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4231719Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4231802Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4232033Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4232131Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4232359Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4232454Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4232534Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4232599Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4232724Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4233076Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4233158Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4233388Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4233483Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4233716Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4233815Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4233897Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4233963Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4234086Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4234452Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4234534Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4234777Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4234874Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4235106Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4235203Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4235436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4235532Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4235764Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4235860Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4235942Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4236007Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4236119Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4236474Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4236558Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4236816Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4236910Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4237140Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4237263Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4237347Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4237414Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4237523Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4237881Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4237964Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4238195Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4238291Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4238536Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4238660Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4238891Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4238989Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4239239Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4239337Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4239418Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4239482Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4239594Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4239946Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4240030Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4240260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4240356Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4240584Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4240683Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4240912Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4241008Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4241240Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4241339Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4241420Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4241498Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4241609Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4241963Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4242048Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4242280Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4242378Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4242609Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4242706Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4242945Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4243054Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4243282Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4243381Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4243477Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4243543Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4243652Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4244009Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4244093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4244323Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4244422Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4244652Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4244751Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4244833Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4244900Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4245009Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4245366Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4245447Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4245679Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4245797Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4246027Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4246124Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4246355Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4246451Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4246678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4246809Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4246892Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4246958Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4247091Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4247461Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4247542Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4247773Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4247883Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4248115Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4248210Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4248439Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4248536Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4248766Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4248860Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4248944Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4249009Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4249119Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4249473Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4249555Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4249786Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4249881Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4250110Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4250223Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4251794Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4251891Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4252122Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4252216Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4252300Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4252365Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4252476Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4252848Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4252941Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4253175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4253270Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4253498Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4253607Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4253842Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4253941Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4254172Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4254268Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4254349Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4254414Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4254524Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4254889Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4254975Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4255207Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4255303Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4255534Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4255629Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4255723Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.4255795Z Traceback (most recent call last): 2025-12-04T09:54:35.4255941Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.4256026Z self.common( 2025-12-04T09:54:35.4256126Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.4256175Z return func(*args, **kwds) 2025-12-04T09:54:35.4256317Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.4256357Z check_model( 2025-12-04T09:54:35.4256488Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.4256532Z assert_equal_fn( 2025-12-04T09:54:35.4256686Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.4256804Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.4256986Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.4257068Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.4257131Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4257154Z 2025-12-04T09:54:35.4257206Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4257315Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4257421Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4257424Z 2025-12-04T09:54:35.4257474Z The failure occurred for item [2] 2025-12-04T09:54:35.4257476Z 2025-12-04T09:54:35.4257559Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4257731Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4257750Z 2025-12-04T09:54:35.4257852Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4257936Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4258003Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4258114Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4258473Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4258555Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4258793Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4258892Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4259125Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4259224Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4259455Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4259551Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4259781Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4259879Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4259960Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4260044Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4260155Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4260540Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4260621Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4260852Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4260947Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4261179Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4261275Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4261371Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4261437Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4261547Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4261901Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4261982Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4262233Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4262331Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4262566Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4262662Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4262895Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4262990Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4263222Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4263318Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4263400Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4263464Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4263577Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4263933Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4264016Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4264249Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4264364Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4264597Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4264710Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4264792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4264855Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4264964Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4265320Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4265402Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4265633Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4265744Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4265973Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4266069Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4266149Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4266214Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4266336Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4266693Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4266810Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4267043Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4267140Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4267370Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4267467Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4267700Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4267798Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4268031Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4268127Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4268209Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4268276Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4268385Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4268742Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4268843Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4269094Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4269191Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4269421Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4269518Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4269749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4269847Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4270079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4270193Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4270274Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4270340Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4270448Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4270818Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4270902Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4271134Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4271232Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4271462Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4271557Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4271787Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4271883Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4272114Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4272211Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4272293Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4272358Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4272467Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4272822Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4272905Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4273135Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4273260Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4273491Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4273586Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4273815Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4273911Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4274142Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4274239Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4274321Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4274399Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4274508Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4274863Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4274944Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4275187Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4275284Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4275515Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4275611Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4275841Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4275936Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4276166Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4276262Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4276345Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4276408Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4276520Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4276925Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4277008Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4277239Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4277335Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4277585Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4277699Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4277782Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4277846Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4277956Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4278311Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4278395Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4278629Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4278729Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4278989Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4279087Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4279316Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4279413Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4279659Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4279757Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4279842Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4279907Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4280017Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4280369Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4280451Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4280681Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4280779Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4281010Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4281109Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4281340Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4281435Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4281664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4281761Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4281859Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4281924Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4282048Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4282404Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4282486Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4282717Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4282813Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4283044Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4283152Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4283383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4283479Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4283711Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4283807Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4283902Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4283966Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4284076Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4284433Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4284516Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4284748Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4284845Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4285078Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4285174Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4285257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4285322Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4285433Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4285788Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4285870Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4286102Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4286212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4286456Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4286553Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4286818Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4286915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4287146Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4287242Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4287326Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4287392Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4287522Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4287877Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4287959Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4288190Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4288301Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4288532Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4288632Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4288862Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4288958Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4289187Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4289284Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4289367Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4289431Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4289541Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4289900Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4289983Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4290215Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4290312Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4290542Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4290653Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4290752Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4290816Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4290923Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4291286Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4291367Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4291599Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4291696Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4291937Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4292034Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4292117Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4292184Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4292294Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4292663Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4292747Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4292982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4293078Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4293308Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4293403Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4293634Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4293729Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4293961Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4294060Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4294142Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4294206Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4294316Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4294673Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4294757Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4295003Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4295120Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4295350Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4295445Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4295676Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4295770Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4296002Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4296098Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4296192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4296256Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4296366Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4296721Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4296840Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4297089Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4297187Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4297422Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4297517Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4297599Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4297661Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4297770Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4298123Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4298207Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4298438Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4298535Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4298764Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4298860Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4298941Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4299008Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4299117Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4299492Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4299592Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4299829Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4299926Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4300159Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4300255Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4300485Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4300596Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4300826Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4300922Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4301003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4301069Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4301189Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4301545Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4301629Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4301861Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4301958Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4302187Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4302286Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4302516Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4302613Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4302846Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4302941Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4303022Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4303086Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4303194Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4303548Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4303641Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4303887Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4303982Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4304212Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4304308Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4304539Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4304636Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4304867Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4304975Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4305057Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4305121Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4305229Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4305601Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4305683Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4305914Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4306014Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4306243Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4306338Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4306419Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4306482Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4306592Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4306985Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4307070Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4307308Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4307404Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4307636Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4307733Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4307814Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4307895Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4308003Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4308376Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4308459Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4308694Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4308795Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4309027Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4309127Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4309374Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4309472Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4309707Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4309805Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4309888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4309966Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4310078Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4310432Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4310518Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4310751Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4310852Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4311082Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4311181Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4311262Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4311329Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4311441Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4311795Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4311880Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4312113Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4312232Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4312462Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4312574Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4312656Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4312721Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4312829Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4313187Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4313270Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4313504Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4313613Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4313848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4313946Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4314032Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4314095Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4314216Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4314576Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4314660Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4314894Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4314989Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4315223Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4315319Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4315401Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4315469Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4315579Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4315939Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4316028Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4316260Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4316359Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4316589Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4316698Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4316982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4317079Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4317307Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4317403Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4317489Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4317554Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4317667Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4318044Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4318128Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4318358Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4318454Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4318698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4318801Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4318883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4318952Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4319063Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4319426Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4319510Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4319741Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4319840Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4320071Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4320170Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4320402Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4320498Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4320728Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4320830Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4320929Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4320996Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4321120Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4321479Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4321561Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4321792Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4321890Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4322122Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4322223Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4322464Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4322562Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4322792Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4322889Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4322983Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4323052Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4323164Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4323528Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4323610Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4323844Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4323942Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4324177Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4324278Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4324513Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4324613Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4324846Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4324945Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4325029Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4325097Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4325208Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4325578Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4325677Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4325910Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4326008Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4326241Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4326337Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4326421Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4326488Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4326613Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4327002Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4327085Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4327320Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4327439Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4327674Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4327773Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4328004Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4328099Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4328330Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4328426Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4328510Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4328576Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4328689Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4329049Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4329133Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4329367Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4329464Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4329697Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4329810Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4330057Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4330153Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4330383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4330478Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4330563Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4330629Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4330742Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4331114Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4331201Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4331433Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4331533Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4331764Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4331876Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4332112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4332212Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4332441Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4332538Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4332623Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4332689Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4332801Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4333158Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4333244Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4333475Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4333572Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4333800Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4333897Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4334127Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4337035Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4337315Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4337413Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4337496Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4337561Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4337671Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4338032Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4338118Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4338368Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4338465Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4338694Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4338790Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4338872Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4338953Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4339065Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4339421Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4339505Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4339738Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4339834Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4340064Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4340161Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4340391Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4340488Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4340717Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4340811Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4340902Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.4340959Z Traceback (most recent call last): 2025-12-04T09:54:35.4341106Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.4341151Z self.common( 2025-12-04T09:54:35.4341251Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.4341319Z return func(*args, **kwds) 2025-12-04T09:54:35.4341461Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.4341515Z check_model( 2025-12-04T09:54:35.4341645Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.4341689Z assert_equal_fn( 2025-12-04T09:54:35.4341843Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.4341912Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.4342091Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.4342172Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.4342234Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4342238Z 2025-12-04T09:54:35.4342290Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4342409Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4342537Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4342542Z 2025-12-04T09:54:35.4342595Z The failure occurred for item [2] 2025-12-04T09:54:35.4342598Z 2025-12-04T09:54:35.4342680Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4342851Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4342854Z 2025-12-04T09:54:35.4342953Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4343036Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4343115Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4343226Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4343583Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4343665Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4343901Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4343998Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4344229Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4344326Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4344556Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4344652Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4344881Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4344976Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4345056Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4345120Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4345230Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4345600Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4345694Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4345925Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4346020Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4346250Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4346345Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4346427Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4346492Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4346603Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4347026Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4347109Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4347339Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4347452Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4347685Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4347781Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4348011Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4348105Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4348333Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4348427Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4348509Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4348573Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4348683Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4349040Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4349122Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4349352Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4349448Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4349677Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4349793Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4349876Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4349956Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4350066Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4350418Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4350500Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4350730Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4350827Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4351068Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4351166Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4351245Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4351311Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4351419Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4351787Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4351870Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4352101Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4352198Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4352427Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4352523Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4352750Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4352846Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4353074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4353171Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4353251Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4353317Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4353425Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4353780Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4353862Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4354108Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4354214Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4354444Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4354539Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4354769Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4354865Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4355095Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4355191Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4355273Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4355349Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4355458Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4355809Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4355890Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4356131Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4356227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4356458Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4356555Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4356848Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4356944Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4357175Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4357271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4357352Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4357417Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4357528Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4357882Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4357963Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4358193Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4358289Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4358538Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4358649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4358880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4358974Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4359204Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4359299Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4359381Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4359446Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4359556Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4359931Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4360012Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4360242Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4360336Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4360579Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4360675Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4360907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4361001Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4361229Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4361323Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4361404Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4361468Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4361577Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4361935Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4362018Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4362247Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4362342Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4362571Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4362666Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4362761Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4362847Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4362958Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4363311Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4363393Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4363623Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4363720Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4363950Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4364058Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4364287Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4364383Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4364613Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4364709Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4364802Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4364868Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4364977Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4365333Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4365415Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4365646Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4365742Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4365971Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4366068Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4366297Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4366393Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4366621Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4366716Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4366841Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4366908Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4367018Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4367395Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4367491Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4367721Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4367816Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4368045Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4368142Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4368372Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4368483Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4368711Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4368807Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4368888Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4368952Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4369076Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4369434Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4369519Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4369752Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4369850Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4370080Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4370176Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4370257Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4370323Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4370431Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4370789Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4370871Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4371103Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4371199Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4371431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4371539Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4371788Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4371883Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4372114Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4372209Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4372289Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4372355Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4372464Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4372830Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4372912Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4373143Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4373238Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4373478Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4373575Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4373805Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4373901Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4374130Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4374224Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4374307Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4374369Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4374480Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4374836Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4374919Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4375150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4375245Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4375473Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4375570Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4375650Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4375729Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4375838Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4376201Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4376284Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4376515Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4376611Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4376890Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4376987Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4377086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4377152Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4377261Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4377615Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4377697Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4377942Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4378039Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4378269Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4378364Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4378591Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4378687Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4378915Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4379012Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4379092Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4379158Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4379268Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4379624Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4379706Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4379936Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4380054Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4380282Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4380394Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4380623Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4380720Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4380949Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4381048Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4381131Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4381194Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4381305Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4381676Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4381758Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4381991Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4382097Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4382325Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4382422Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4382503Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4382567Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4382675Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4383028Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4383109Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4383338Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4383434Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4383664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4383758Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4383838Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4383903Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4384012Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4384372Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4384471Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4384714Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4384809Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4385039Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4385134Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4385365Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4385461Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4385705Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4385802Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4385883Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4385948Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4386057Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4386427Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4386511Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4386785Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4386881Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4387111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4387205Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4387434Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4387530Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4387763Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4387859Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4387940Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4388006Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4388114Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4388468Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4388549Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4388794Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4388902Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4389134Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4389230Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4389460Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4389554Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4389784Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4389881Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4389964Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4390042Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4390151Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4390503Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4390583Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4390826Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4390922Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4391156Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4391252Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4391333Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4391395Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4391507Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4391866Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4391950Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4392181Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4392278Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4392510Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4392605Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4392687Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4392751Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4392861Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4393230Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4393323Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4393553Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4393649Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4393880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4393978Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4394208Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4394306Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4394545Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4394642Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4394723Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4394786Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4394894Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4395264Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4395348Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4395578Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4395673Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4395901Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4395996Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4396077Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4396140Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4396250Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4396608Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4396688Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4396952Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4397046Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4397275Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4397390Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4397471Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4397553Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4397661Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4398017Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4398098Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4398329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4398424Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4398672Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4398768Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4398849Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4398912Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4399021Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4399387Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4399471Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4399704Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4399800Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4400031Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4400125Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4400207Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4400272Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4400383Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4400741Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4400824Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4401055Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4401153Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4401384Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4401482Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4401720Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4401827Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4402056Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4402151Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4402232Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4402295Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4402403Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4402757Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4402841Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4403079Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4403175Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4403402Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4403497Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4403588Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4403654Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4403764Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4404119Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4404200Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4404429Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4404524Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4404752Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4404847Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4405075Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4405172Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4405400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4405495Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4405576Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4405641Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4405751Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4406118Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4406211Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4406442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4406537Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4406802Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4406897Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4407126Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4407239Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4407467Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4407562Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4407643Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4407707Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4407816Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4408186Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4408271Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4408502Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4408597Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4408826Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4408921Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4409150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4409245Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4409476Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4409571Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4409651Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4409715Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4409823Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4410177Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4410272Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4410519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4410613Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4410843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4410938Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4411019Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4411084Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4411194Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4411566Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4411651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4411881Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4411978Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4412210Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4412316Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4412551Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4412647Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4412879Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4412973Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4413055Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4413119Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4413227Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4413585Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4413674Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4413907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4414004Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4414233Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4414330Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4414561Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4414667Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4414909Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4415004Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4415086Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4415150Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4415261Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4415616Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4415700Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4415943Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4416041Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4416270Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4416365Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4416604Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4416701Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4416974Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4417073Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4417155Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4417221Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4417334Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4417689Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4417771Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4418004Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4418101Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4418333Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4418429Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4418659Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4418753Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4418982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4419095Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4419192Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4419258Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4419367Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4419723Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4419806Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4420036Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4420133Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4420376Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4420473Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4420554Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4420622Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4420730Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4421104Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4421187Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4421423Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4421520Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4421749Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4421844Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4422075Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4422173Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4422401Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4422498Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4422579Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4422645Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4422753Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4423107Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4423189Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4427388Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4434070Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4434303Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4434399Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4434491Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.4434549Z Traceback (most recent call last): 2025-12-04T09:54:35.4434694Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.4434739Z self.common( 2025-12-04T09:54:35.4434839Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.4434889Z return func(*args, **kwds) 2025-12-04T09:54:35.4435031Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.4435087Z check_model( 2025-12-04T09:54:35.4435220Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.4435262Z assert_equal_fn( 2025-12-04T09:54:35.4435417Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.4435486Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.4435663Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.4435763Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.4435828Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4435831Z 2025-12-04T09:54:35.4435881Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4435995Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4436112Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4436115Z 2025-12-04T09:54:35.4436164Z The failure occurred for item [2] 2025-12-04T09:54:35.4436166Z 2025-12-04T09:54:35.4436250Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4436420Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4436423Z 2025-12-04T09:54:35.4436523Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4436610Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4436676Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4436808Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4437166Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4437248Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4437481Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4437580Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4437812Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4437931Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4438160Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4438279Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4438511Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4438608Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4438691Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4438756Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4438867Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4439240Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4439322Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4439554Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4439650Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4439880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4439989Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4440073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4440139Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4440252Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4440609Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4440691Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4440926Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4441024Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4441256Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4441354Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4441585Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4441681Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4441909Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4442006Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4442089Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4442169Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4442280Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4442646Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4442728Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4442957Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4443054Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4443283Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4443378Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4443460Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4443535Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4443644Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4443997Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4444078Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4444321Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4444419Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4444649Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4444745Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4444825Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4444891Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4445001Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4445359Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4445444Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4445678Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4445774Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4446003Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4446100Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4446329Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4446425Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4446667Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4446829Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4446909Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4446976Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4447084Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4447439Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4447522Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4447756Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4447871Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4448100Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4448196Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4448424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4448519Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4448769Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4448867Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4448949Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4449017Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4449125Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4449480Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4449561Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4449792Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4449888Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4450118Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4450213Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4450442Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4450537Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4450765Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4450863Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4450961Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4451040Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4451151Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4451513Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4451596Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4451833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4451930Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4452163Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4452271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4452500Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4452596Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4452828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4452925Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4453016Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4453084Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4453195Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4453554Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4453636Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4453871Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4453967Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4454199Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4454297Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4454530Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4454627Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4454857Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4454952Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4455035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4455100Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4455229Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4455585Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4455679Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4455910Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4456006Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4456238Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4456334Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4456416Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4456481Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4456605Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4457009Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4457093Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4457347Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4457446Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4457681Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4457781Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4458019Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4458116Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4458351Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4458449Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4458533Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4458600Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4458713Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4459078Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4459161Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4459395Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4459494Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4459728Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4459842Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4460089Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4460186Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4460417Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4460514Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4460597Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4460662Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4460774Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4461142Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4461230Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4461463Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4461562Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4461813Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4461913Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4462145Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4462245Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4462478Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4462573Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4462654Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4462720Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4462830Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4463188Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4463273Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4463505Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4463602Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4463833Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4463931Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4464011Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4464088Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4464207Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4464566Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4464651Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4464884Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4464981Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4465211Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4465308Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4465550Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4465646Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4465878Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4465976Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4466069Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4466135Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4466245Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4466602Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4466684Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4466955Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4467051Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4467284Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4467380Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4467615Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4467711Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4467943Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4468043Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4468123Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4468187Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4468296Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4468672Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4468770Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4469006Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4469102Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4469335Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4469432Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4469515Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4469579Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4469705Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4470060Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4470142Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4470375Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4470491Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4470724Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4470821Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4470903Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4470968Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4471077Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4471436Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4471519Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4471752Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4471850Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4472081Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4472178Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4472408Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4472503Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4472734Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4472843Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4472923Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4473001Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4473111Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4473468Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4473549Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4473781Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4473879Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4474124Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4474223Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4474452Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4474548Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4474776Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4474882Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4474965Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4475028Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4475140Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4475500Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4475584Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4475819Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4475916Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4476150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4476248Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4476330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4476394Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4476503Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4476884Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4476966Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4477221Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4477333Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4477569Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4477665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4477747Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4477812Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4477923Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4478283Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4478366Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4478614Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4478709Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4478942Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4479037Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4479282Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4479379Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4479613Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4479710Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4479792Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4479857Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4479967Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4480323Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4480406Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4480640Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4480735Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4480965Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4481060Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4481292Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4481388Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4481630Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4481737Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4481819Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4481883Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4481993Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4482347Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4482430Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4482662Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4482769Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4483000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4483095Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4483327Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4483422Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4483664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4483765Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4483849Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4483913Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4484023Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4484384Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4484466Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4484700Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4484796Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4485028Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4485124Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4485204Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4485267Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4485375Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4485732Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4485825Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4486070Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4486166Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4486397Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4486493Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4486572Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4486637Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4486786Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4487169Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4487252Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4487486Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4487581Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4487815Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4487929Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4488163Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4488261Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4488490Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4488586Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4488666Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4488729Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4488838Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4489196Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4489280Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4489512Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4489607Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4489842Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4489938Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4490020Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4490102Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4490212Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4490582Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4490666Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4490902Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4490997Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4491228Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4491323Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4491405Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4491479Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4491588Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4491941Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4492023Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4492266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4492365Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4492599Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4492696Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4492776Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4492839Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4492949Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4493307Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4493392Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4493629Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4493726Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4493959Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4494055Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4494135Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4494201Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4494311Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4494690Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4494781Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4495015Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4495111Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4495342Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4495438Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4495670Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4495777Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4496009Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4496105Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4496185Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4496249Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4496359Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4496726Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4496847Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4497082Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4497178Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4497411Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4497505Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4497586Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4497653Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4497762Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4498122Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4498205Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4498439Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4498535Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4498768Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4498881Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4499130Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4499225Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4499461Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4499555Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4499638Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4499704Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4499815Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4500192Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4500276Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4500511Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4500608Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4500861Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4500956Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4501191Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4501292Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4501526Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4501622Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4501706Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4501771Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4501885Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4502239Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4502324Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4502555Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4502652Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4502884Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4502982Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4503216Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4503324Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4503576Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4503672Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4503753Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4503816Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4503926Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4504287Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4504373Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4504619Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4504717Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4504948Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4505045Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4505125Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4505202Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4505312Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4505667Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4505750Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4505982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4506078Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4506310Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4506406Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4506636Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4506735Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4507014Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4507111Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4507191Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4507256Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4507372Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4507729Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4507841Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4508075Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4508171Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4508403Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4508499Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4508733Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4508830Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4509074Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4509175Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4509255Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4509320Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4509429Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4509799Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4509883Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4510121Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4510217Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4510453Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4510549Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4510782Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4510879Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4511110Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4511208Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4511289Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4511355Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4511464Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4511824Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4511920Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4512154Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4512262Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4512495Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4512590Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4512825Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4512922Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4513152Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4513249Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4513342Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4513406Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4513515Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4513871Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4513967Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4514200Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4514296Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4514530Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4514629Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4514710Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4514774Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4514883Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4515238Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4515323Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4515558Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4515654Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4515886Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4515981Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4516213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4516322Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4516556Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4516663Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4516746Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4516884Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4516996Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4517353Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4517437Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4517684Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4517783Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4518014Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4518109Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4518190Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4518253Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4518383Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4518738Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4518822Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4519054Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4519150Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4519382Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4519479Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4519570Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.4519624Z Traceback (most recent call last): 2025-12-04T09:54:35.4519766Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.4519810Z self.common( 2025-12-04T09:54:35.4519907Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.4519956Z return func(*args, **kwds) 2025-12-04T09:54:35.4520099Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 725, in check_model_gpu 2025-12-04T09:54:35.4520142Z check_model( 2025-12-04T09:54:35.4520271Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.4520313Z assert_equal_fn( 2025-12-04T09:54:35.4520468Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.4520539Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.4520744Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.4520840Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.4520903Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4520906Z 2025-12-04T09:54:35.4520957Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4521066Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4521172Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4521174Z 2025-12-04T09:54:35.4521225Z The failure occurred for item [2] 2025-12-04T09:54:35.4521228Z 2025-12-04T09:54:35.4521309Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4521482Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4521485Z 2025-12-04T09:54:35.4521588Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4521673Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4521752Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4521863Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4522217Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4522302Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4522547Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4522648Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4522885Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4522983Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4523213Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4523310Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4523540Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4523636Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4523718Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4523781Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4523895Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4524250Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4524332Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4524564Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4524662Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4524904Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4525012Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4525093Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4525159Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4525268Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4525625Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4525708Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4525943Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4526051Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4526282Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4526378Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4526608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4526704Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4526987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4527086Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4527169Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4527234Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4527344Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4527708Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4527791Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4528028Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4528125Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4528361Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4528458Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4528539Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4528603Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4528713Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4529070Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4529169Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4529416Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4529512Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4529744Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4529839Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4529921Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4529986Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4530098Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4530470Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4530554Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4530787Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4530883Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4531115Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4531220Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4531456Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4531554Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4531787Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4531882Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4531964Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4532028Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4532139Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4532497Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4532583Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4532817Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4532915Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4533145Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4533241Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4533474Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4533580Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4533825Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4533921Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4534003Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4534068Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4534180Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4534538Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4534621Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4534869Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4534967Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4535201Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4535299Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4535555Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4535652Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4535884Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4535981Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4536063Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4536127Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4536396Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4536797Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4536880Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4537115Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4537214Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4537443Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4537539Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4537769Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4537867Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4538101Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4538222Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4538318Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4538384Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4538492Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4538847Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4538929Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4539161Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4539259Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4539505Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4539602Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4539830Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4539926Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4540171Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4540271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4540352Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4540420Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4540530Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4540890Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4540972Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4541209Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4541307Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4541538Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4541637Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4541717Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4541783Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4541892Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4542252Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4542344Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4542577Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4542685Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4542919Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4543015Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4543246Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4543342Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4543574Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4543672Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4543764Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4543829Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4543939Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4544295Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4544388Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4544622Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4544719Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4544952Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4545047Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4545278Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4545373Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4545608Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4545705Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4545786Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4545852Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4545962Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4546316Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4546398Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4546631Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4546773Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4547005Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4547132Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4547362Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4547458Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4547687Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4547783Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4547868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4547930Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4548056Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4548412Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4548494Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4548728Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4548838Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4549071Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4549168Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4549251Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4549316Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4549425Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4549781Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4549864Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4550097Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4550195Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4550431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4550528Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4550757Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4550853Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4551086Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4551195Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4551318Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4553963Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4554079Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4554438Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4554521Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4554757Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4554857Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4555112Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4555209Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4555441Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4555537Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4555767Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4555876Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4555961Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4556024Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4556137Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4556493Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4556575Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4556875Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4556971Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4557202Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4557300Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4557380Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4557443Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4557552Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4557907Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4557991Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4558246Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4558358Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4558587Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4558683Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4558763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4558828Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4558937Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4559290Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4559373Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4559621Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4559718Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4559947Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4560043Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4560287Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4560385Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4560618Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4560716Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4560798Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4560863Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4560972Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4561327Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4561410Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4561641Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4561738Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4561968Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4562064Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4562294Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4562391Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4562633Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4562741Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4562822Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4562886Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4562995Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4563350Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4563432Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4563664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4563772Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4564000Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4564095Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4564176Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4564240Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4564348Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4564712Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4564796Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4565027Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4565122Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4565360Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4565459Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4565541Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4565608Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4565716Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4566071Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4566153Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4566383Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4566478Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4566708Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4566856Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4567102Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4567197Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4567426Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4567520Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4567602Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4567666Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4567779Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4568151Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4568236Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4568468Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4568565Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4568796Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4568910Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4569141Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4569238Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4569466Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4569560Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4569642Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4569705Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4569814Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4570170Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4570255Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4570484Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4570580Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4570808Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4570903Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4571133Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4574731Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4581814Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4581909Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4581990Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4582053Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4582163Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4582520Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4582603Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4582857Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4582953Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4583183Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4583279Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4583359Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4583444Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4583552Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4583908Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4583993Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4584221Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4584316Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4584544Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4584640Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4584721Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4584787Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4584896Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4585252Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4585332Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4585562Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4585658Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4585905Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4586017Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4586246Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4586341Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4586570Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4586665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4586784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4586848Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4586956Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4587332Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4587414Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4587647Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4587742Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4587989Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4588085Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4588168Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4588230Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4588339Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4588691Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4588772Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4589005Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4589101Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4589332Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4589427Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4589508Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4589569Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4589680Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4590037Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4590137Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4590384Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4590480Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4590708Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4590803Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4590884Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4590949Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4591056Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4591422Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4591506Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4591736Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4591832Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4592072Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4592170Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4592251Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4592319Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4592429Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4592787Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4592868Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4593102Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4593198Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4593431Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4593528Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4593760Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4593856Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4594087Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4594184Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4594265Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4594345Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4594454Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4594822Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4594903Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4595134Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4595228Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4595459Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4595554Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4595651Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4595717Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4595826Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4596181Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4596263Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4596519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4596616Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4596913Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4597007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4597235Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4597328Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4597559Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4597653Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4597735Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4597803Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4597913Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4598264Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4598345Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4598575Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4598670Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4598920Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4599032Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4599263Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4599357Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4599586Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4599681Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4599763Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4599828Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4599937Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4600306Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4600390Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4600621Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4600718Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4600963Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4601061Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4601293Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4601389Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4601624Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4601720Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4601801Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4601865Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4601975Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4602331Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4602414Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4602644Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4602740Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4602972Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4603083Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4603163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4603241Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4603350Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4603705Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4603787Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4604019Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4604116Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4604345Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4604455Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4604685Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4604781Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4605010Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4605119Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4605200Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4605266Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4605375Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4605731Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4605812Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4606042Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4606138Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4606372Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4606469Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4606701Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4606864Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4607093Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4607192Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4607273Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4607338Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4607465Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4607818Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4607913Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4608147Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4608241Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4608472Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4608567Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4608813Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4608910Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4609139Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4609234Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4609314Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4609379Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4609501Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4609859Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4609942Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4610173Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4610268Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4610497Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4610592Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4610824Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4610919Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4611151Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4611246Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4611327Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4611391Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4611500Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4611857Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4611962Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4612196Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4612291Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4612519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4612614Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4612696Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4612760Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4612870Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4613243Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4613328Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4613558Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4613654Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4613894Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4613991Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4614224Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4614320Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4614552Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4614650Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4614731Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4614794Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4614907Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4615264Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4615346Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4615579Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4615675Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4615907Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4616003Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4616097Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4616160Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4616282Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4616636Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4616717Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4617132Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4617230Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4617461Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4617556Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4617658Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4617723Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4617833Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4618190Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4618286Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4618519Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4618617Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4618851Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4618947Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4619176Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4619271Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4619503Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4619600Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4619693Z _______________ GPUTests.test_var_mean_tile_reduction_True_cuda ________________ 2025-12-04T09:54:35.4619748Z Traceback (most recent call last): 2025-12-04T09:54:35.4619895Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 5770, in test_var_mean 2025-12-04T09:54:35.4619938Z self.common( 2025-12-04T09:54:35.4620037Z File "/opt/conda/envs/py_3.10/lib/python3.10/contextlib.py", line 79, in inner 2025-12-04T09:54:35.4620085Z return func(*args, **kwds) 2025-12-04T09:54:35.4620228Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 692, in check_model_gpu 2025-12-04T09:54:35.4620269Z check_model( 2025-12-04T09:54:35.4620399Z File "/var/lib/jenkins/pytorch/test/inductor/test_torchinductor.py", line 566, in check_model 2025-12-04T09:54:35.4620442Z assert_equal_fn( 2025-12-04T09:54:35.4620613Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_dynamo/test_case.py", line 113, in assertEqual 2025-12-04T09:54:35.4620681Z return super().assertEqual(x, y, *args, **kwargs) 2025-12-04T09:54:35.4620874Z File "/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/testing/_internal/common_utils.py", line 4284, in assertEqual 2025-12-04T09:54:35.4620955Z raise error_metas.pop()[0].to_error( # type: ignore[index] 2025-12-04T09:54:35.4621015Z AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4621018Z 2025-12-04T09:54:35.4621070Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4621188Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4621302Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4621304Z 2025-12-04T09:54:35.4621356Z The failure occurred for item [2] 2025-12-04T09:54:35.4621358Z 2025-12-04T09:54:35.4621441Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4621613Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4621616Z 2025-12-04T09:54:35.4621725Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4621809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4621874Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4621986Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4622361Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4622444Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4622682Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4622783Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4623015Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4623111Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4623342Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4623438Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4623670Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4623766Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4623850Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4623914Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4624025Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4624380Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4624464Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4624698Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4624809Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4625051Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4625145Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4625227Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4625291Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4625399Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4625754Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4625836Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4626078Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4626176Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4626403Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4626499Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4626773Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4626871Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4627099Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4627197Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4627278Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4627340Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4627450Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4627807Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4627889Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4628119Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4628217Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4628445Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4628540Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4628620Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4628683Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4628792Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4629145Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4629268Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4629500Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4629596Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4629828Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4629923Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4630005Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4630075Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4630183Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4630553Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4630635Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4630865Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4630960Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4631203Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4631300Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4631533Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4631628Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4631860Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4631955Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4632039Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4632106Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4632217Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4632576Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4632658Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4632891Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4632986Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4633218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4633326Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4633557Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4633664Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4633893Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4633988Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4634070Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4634134Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4634244Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4634601Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4634695Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4634928Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4635024Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4635254Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4635358Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4635588Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4635684Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4635914Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4636008Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4636090Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4636153Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4636263Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4636617Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4636701Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4636983Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4637081Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4637310Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4637404Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4637634Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4637744Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4637975Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4638083Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4638165Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4638228Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4638338Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4638694Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4638777Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4639022Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4639120Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4639347Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4639442Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4639674Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4639784Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4640016Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4640112Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4640194Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4640257Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4640367Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4640720Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4640803Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4641034Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4641130Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4641360Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4641456Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4641536Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4641601Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4641708Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4642068Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4642173Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4642404Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4642500Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4642729Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4642825Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4643054Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4643150Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4643389Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4643486Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4643568Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4643633Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4643742Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4644112Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4644195Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4644430Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4644528Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4644757Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4644853Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4645081Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4645179Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4645410Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4645510Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4645590Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4645656Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4645764Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4646120Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4646203Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4646454Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4646562Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4646834Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4646930Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4647159Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4647254Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4647483Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4647583Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4647663Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4647741Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4647850Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4648204Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4648284Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4648532Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4648629Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4648863Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4648958Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4649040Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4649104Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4649213Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4649574Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4649657Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4649891Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4649987Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4650218Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4650313Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4650543Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4650639Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4650883Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4650992Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4651073Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4651138Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4651248Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4651602Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4651686Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4651923Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4652029Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4652263Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4652357Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4652585Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4652681Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4652922Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4653018Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4653102Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4653165Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4653274Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4653627Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4653709Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4653940Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4654037Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4654270Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4654367Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4654448Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4654510Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4654621Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4654976Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4655070Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4655312Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4655407Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4655634Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4655729Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4655809Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4655875Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4655983Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4656350Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4656433Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4656665Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4656792Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4657038Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4657135Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4657366Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4657464Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4657692Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4657787Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4657868Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4657932Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4658042Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4658395Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4658478Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4658709Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4658803Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4659032Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4659126Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4659356Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4659469Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4659712Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4659808Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4659889Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4659954Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4660062Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4660419Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4660502Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4660746Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4660841Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4661072Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4661166Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4661248Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4661334Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4661445Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4661796Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4661880Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4662111Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4662206Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4662436Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4662531Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4662613Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4662679Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4662788Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4663141Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4663223Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4663453Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4663549Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4663791Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4663899Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4664130Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4664227Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4664458Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4664554Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4664637Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4664702Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4664813Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4665178Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4665261Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4665491Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4665589Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4665830Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4665928Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4666160Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4666258Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4666490Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4666586Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4666668Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4666734Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4666902Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4667259Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4667343Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4667573Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4667672Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4667900Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4668025Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4668255Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4668379Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4668607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4668703Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4668784Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4668849Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4668958Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4669315Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4669422Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4669656Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4669752Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4669982Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4670101Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4670183Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4670249Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4670356Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4670715Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4670797Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4671030Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4671127Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4671357Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4671453Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4671538Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4671603Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4671713Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4672073Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4672155Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4672389Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4672499Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4672744Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4672841Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4673073Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4673167Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4673400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4673498Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4673581Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4673657Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4673771Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4674126Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4674207Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4674452Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4674549Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4674779Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4674876Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4674959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4675022Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4675133Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4675491Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4675576Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4675806Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4675905Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4676134Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4676230Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4676310Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4676375Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4676486Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4676906Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4677004Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4677235Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4677331Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4677561Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4677658Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4677738Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4677802Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4677910Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4678289Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4678370Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4678604Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4678716Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4678948Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4679046Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4679134Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4679198Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4679309Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4679667Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4679749Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4679983Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4680079Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4680312Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4680408Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4680638Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4680732Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4680965Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4681085Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4681168Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4681243Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4681353Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4681708Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4681792Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4682025Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4682122Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4682354Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4682460Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4682543Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4682608Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4682719Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4683087Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4683170Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4683400Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4683499Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4683728Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4683824Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4684053Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4684150Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4684380Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4684477Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4684560Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4684625Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4684737Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4685096Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4685179Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4685408Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4685516Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4685757Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4685854Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4686082Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4686178Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4686407Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4686503Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4686583Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4686661Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4686811Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4687170Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4687252Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4687499Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4687597Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4687829Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4687926Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4688155Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4688251Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4688479Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4688576Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4688658Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4688722Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4688832Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4689186Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4689269Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4689500Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4689597Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4689843Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4689954Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4690035Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4690102Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4690210Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4690565Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4690648Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4690880Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4690976Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4691219Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4691315Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4691545Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4691640Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4691884Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4691982Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4692066Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4692133Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4692242Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4692598Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4692679Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4692911Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4693007Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4693238Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4693333Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4693562Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4693657Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4693886Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4693981Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4694078Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4694142Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4694270Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4694626Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4694707Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4694939Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4695035Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4695266Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4695375Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4695607Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4695702Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4695932Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4696027Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4696122Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4696187Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4696297Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4696652Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4696734Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4697007Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4697104Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4697337Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4697433Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4697664Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4697758Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4697987Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4698082Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4698163Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4698227Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4698336Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4698730Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4698832Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4699061Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4699157Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4699385Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4699481Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4699565Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4699630Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4699755Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4700110Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4700193Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4700424Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4700535Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4700764Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4700861Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4701089Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4701184Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4701412Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4701507Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4701588Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4701652Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4701759Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4702121Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4702203Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4702432Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4702527Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4702756Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4702866Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4702959Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4703023Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4703129Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4703484Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4703565Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4703798Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4703893Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4704134Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4704231Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4704312Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4704377Z stats [('calls_captured', 12), ('unique_graphs', 2)] 2025-12-04T09:54:35.4704485Z aot_autograd [('total', 2), ('autograd_cache_miss', 2), ('autograd_cache_saved', 2), ('ok', 2)] 2025-12-04T09:54:35.4704851Z inductor [('triton_bundler_save_kernel', 80), ('benchmarking.InductorBenchmarker.benchmark', 10), ('benchmarking.InductorBenchmarker.benchmark_gpu', 10), ('async_compile_cache_miss', 4), ('fxgraph_cache_miss', 2), ('triton_bundler_save_static_autotuner', 2)] 2025-12-04T09:54:35.4704935Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4705168Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4705264Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4705495Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4705590Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4705823Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4705918Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4706150Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4706248Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4706330Z ----------------------------- Captured stdout call ----------------------------- 2025-12-04T09:54:35.4706392Z stats [('calls_captured', 6), ('unique_graphs', 1)] 2025-12-04T09:54:35.4706502Z aot_autograd [('total', 1), ('autograd_cache_miss', 1), ('autograd_cache_saved', 1), ('ok', 1)] 2025-12-04T09:54:35.4706893Z inductor [('triton_bundler_save_kernel', 40), ('benchmarking.InductorBenchmarker.benchmark', 5), ('benchmarking.InductorBenchmarker.benchmark_gpu', 5), ('async_compile_cache_miss', 2), ('fxgraph_cache_miss', 1), ('triton_bundler_save_static_autotuner', 1)] 2025-12-04T09:54:35.4706976Z ----------------------------- Captured stderr call ----------------------------- 2025-12-04T09:54:35.4707225Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__getstate__, please add missing op schema 2025-12-04T09:54:35.4707339Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4707570Z /opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/_inductor/ops_handler.py:772: UserWarning: undefined OpHandler.__setstate__, please add missing op schema 2025-12-04T09:54:35.4707665Z warnings.warn(f"undefined OpHandler.{name}, please add missing op schema") 2025-12-04T09:54:35.4707934Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_compile_subprocess/inductor.test_compile_subprocess-90b2f4794b3bee89.xml - 2025-12-04T09:54:35.4708001Z =========================== short test summary info ============================ 2025-12-04T09:54:35.4708288Z FAILED [0.4716s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4708430Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4708759Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4708916Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4709029Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4713285Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4713489Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4713493Z 2025-12-04T09:54:35.4713580Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4713748Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4713750Z 2025-12-04T09:54:35.4713850Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4714056Z FAILED [1.0047s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4714059Z 2025-12-04T09:54:35.4714111Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4714220Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4714327Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4714330Z 2025-12-04T09:54:35.4714380Z The failure occurred for item [2] 2025-12-04T09:54:35.4714384Z 2025-12-04T09:54:35.4714467Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4714638Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4714642Z 2025-12-04T09:54:35.4714739Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4714991Z FAILED [0.8151s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4715127Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4715432Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4720173Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4720284Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4720391Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4720590Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4720593Z 2025-12-04T09:54:35.4720676Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4720842Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4720844Z 2025-12-04T09:54:35.4720941Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4721205Z FAILED [0.8134s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4721339Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4721642Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4721807Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4721917Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4722026Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4722220Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4722223Z 2025-12-04T09:54:35.4722304Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4722466Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4722469Z 2025-12-04T09:54:35.4722562Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4722810Z FAILED [0.8607s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4722942Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4723241Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4723388Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4723497Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4723602Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4723816Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4723832Z 2025-12-04T09:54:35.4723913Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4724074Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4724077Z 2025-12-04T09:54:35.4724170Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4724415Z FAILED [0.8104s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4724547Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4724845Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4725006Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4725114Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4725219Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4725412Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4725416Z 2025-12-04T09:54:35.4725507Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4725670Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4725679Z 2025-12-04T09:54:35.4725773Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4726022Z FAILED [0.8767s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4726152Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4726451Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4726597Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4726706Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4726848Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4727044Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4727046Z 2025-12-04T09:54:35.4727125Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4727287Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4727290Z 2025-12-04T09:54:35.4727387Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4727650Z FAILED [0.8423s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4727797Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4728093Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4728241Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4728348Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4728455Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4728649Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4728652Z 2025-12-04T09:54:35.4728749Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4728911Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4728914Z 2025-12-04T09:54:35.4729007Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4729253Z FAILED [0.8598s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4729399Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4729700Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4729850Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4729959Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4730063Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4730258Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4730262Z 2025-12-04T09:54:35.4730342Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4730505Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4730509Z 2025-12-04T09:54:35.4730604Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4730849Z FAILED [0.7963s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4730980Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4731278Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4731438Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4731559Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4731665Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4731856Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4731859Z 2025-12-04T09:54:35.4731939Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4732098Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4732102Z 2025-12-04T09:54:35.4732195Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4732442Z FAILED [0.5521s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4732584Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4732884Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4733030Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4733157Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4733263Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4733458Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4733462Z 2025-12-04T09:54:35.4733540Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4733703Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4733707Z 2025-12-04T09:54:35.4733800Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4734045Z FAILED [0.6172s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4734180Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4734481Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4734629Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4734737Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4734843Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4735037Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4735051Z 2025-12-04T09:54:35.4735131Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4735292Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4735307Z 2025-12-04T09:54:35.4735401Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4735647Z FAILED [0.7064s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4735777Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4736077Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4736224Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4736346Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4736450Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4736643Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4736646Z 2025-12-04T09:54:35.4736724Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4736935Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4736963Z 2025-12-04T09:54:35.4737058Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4737304Z FAILED [0.2027s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4737436Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4737732Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4737878Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4737986Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4738094Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4738289Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4738292Z 2025-12-04T09:54:35.4738372Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4738533Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4738536Z 2025-12-04T09:54:35.4738628Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4738875Z FAILED [0.1982s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4739004Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4739331Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4739492Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4739600Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4739703Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4739899Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4739902Z 2025-12-04T09:54:35.4739981Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4740143Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4740163Z 2025-12-04T09:54:35.4740258Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4740503Z FAILED [0.2091s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4740634Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4740943Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4741093Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4741203Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4741309Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4741506Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4741508Z 2025-12-04T09:54:35.4741588Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4741752Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4741755Z 2025-12-04T09:54:35.4741849Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4742096Z FAILED [0.7491s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4742227Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4742526Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4742674Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4742786Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4742903Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4743109Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4743112Z 2025-12-04T09:54:35.4743191Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4743354Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4743356Z 2025-12-04T09:54:35.4743450Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4743701Z FAILED [0.7017s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4743835Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4744145Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4744295Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4744403Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4744509Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4744714Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4744719Z 2025-12-04T09:54:35.4744799Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4744964Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4744967Z 2025-12-04T09:54:35.4745060Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4745307Z FAILED [0.8054s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4745437Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4745738Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4745887Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4745999Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4746104Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4746300Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4746302Z 2025-12-04T09:54:35.4746381Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4746547Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4746560Z 2025-12-04T09:54:35.4746657Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4746944Z FAILED [0.6032s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4747090Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4747392Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4747542Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4747649Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4747757Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4747975Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4747977Z 2025-12-04T09:54:35.4748058Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4748222Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4748226Z 2025-12-04T09:54:35.4748319Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4748584Z FAILED [0.6044s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4748716Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4749018Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4749170Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4749280Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4749385Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4749582Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4749585Z 2025-12-04T09:54:35.4749664Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4749830Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4749832Z 2025-12-04T09:54:35.4749927Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4750174Z FAILED [0.6036s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4750305Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4750606Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4750775Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4750893Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4750998Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4751192Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4751194Z 2025-12-04T09:54:35.4751274Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4751436Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4751441Z 2025-12-04T09:54:35.4751534Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4751793Z FAILED [0.8011s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4751924Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4752224Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4752380Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4752491Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4752595Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4752794Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4752796Z 2025-12-04T09:54:35.4752876Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4753043Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4753045Z 2025-12-04T09:54:35.4753139Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4753392Z FAILED [0.6040s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4753526Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4753833Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4753984Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4754096Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4754203Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4754399Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4754413Z 2025-12-04T09:54:35.4754507Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4754672Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4754677Z 2025-12-04T09:54:35.4754770Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4755017Z FAILED [0.8472s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4755149Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4755455Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4755618Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4755732Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4755837Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4756037Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4756040Z 2025-12-04T09:54:35.4756120Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4756304Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4756307Z 2025-12-04T09:54:35.4756401Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4756650Z FAILED [0.7606s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4756818Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4757120Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4757273Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4757384Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4757493Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4757689Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4757691Z 2025-12-04T09:54:35.4757772Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4757936Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4757941Z 2025-12-04T09:54:35.4758035Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4758286Z FAILED [0.5596s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4758442Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4758763Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4758913Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4759022Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4759129Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4759326Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4759329Z 2025-12-04T09:54:35.4759408Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4759596Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4759598Z 2025-12-04T09:54:35.4759694Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4759943Z FAILED [0.6548s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4760075Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4760388Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4760541Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4760653Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4760762Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4760958Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4760961Z 2025-12-04T09:54:35.4761044Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4761209Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4761214Z 2025-12-04T09:54:35.4761309Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4761560Z FAILED [0.8677s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4761693Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4761998Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4762147Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4762271Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4762396Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4762592Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4762594Z 2025-12-04T09:54:35.4762675Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4762841Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4762843Z 2025-12-04T09:54:35.4762943Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4763194Z FAILED [0.8361s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4763337Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4763637Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4763790Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4763909Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4764031Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4764225Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4764230Z 2025-12-04T09:54:35.4764311Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4764474Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4764479Z 2025-12-04T09:54:35.4764572Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4764822Z FAILED [0.8547s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4764954Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4765264Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4765414Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4765524Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4765631Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4765831Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4765833Z 2025-12-04T09:54:35.4765915Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4766088Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4766101Z 2025-12-04T09:54:35.4766197Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4766448Z FAILED [0.8609s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4766580Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4766918Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4767065Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4767175Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4767299Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4767498Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4767500Z 2025-12-04T09:54:35.4767582Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4767746Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4767749Z 2025-12-04T09:54:35.4767858Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4768109Z FAILED [0.6521s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4768243Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4768545Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4768692Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4768803Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4768910Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4769105Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4769108Z 2025-12-04T09:54:35.4769188Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4769351Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4769353Z 2025-12-04T09:54:35.4769449Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4769697Z FAILED [0.2129s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4769832Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4770155Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4770334Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4770443Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4770551Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4770748Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4770751Z 2025-12-04T09:54:35.4770834Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4770996Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4771000Z 2025-12-04T09:54:35.4771108Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4771357Z FAILED [0.2063s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4771488Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4771803Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4771955Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4772065Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4772171Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4772368Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4772370Z 2025-12-04T09:54:35.4772450Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4772615Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4772618Z 2025-12-04T09:54:35.4772715Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4772964Z FAILED [0.2141s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4773099Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4773400Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4773553Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4773662Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4773769Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4773974Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4773988Z 2025-12-04T09:54:35.4774071Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4774236Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4774239Z 2025-12-04T09:54:35.4774332Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4774579Z FAILED [0.2101s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4774712Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4775026Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4775176Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4775287Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4775392Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4775598Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4775600Z 2025-12-04T09:54:35.4775682Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4775845Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4775849Z 2025-12-04T09:54:35.4775944Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4776195Z FAILED [0.6039s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4776329Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4776631Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4776817Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4776929Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4777037Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4777231Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4777235Z 2025-12-04T09:54:35.4777320Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4777482Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4777486Z 2025-12-04T09:54:35.4777580Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4777861Z FAILED [0.2072s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4778009Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4778313Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4778460Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4778576Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4778683Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4778895Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4778899Z 2025-12-04T09:54:35.4778979Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4779146Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4779148Z 2025-12-04T09:54:35.4779245Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4779492Z FAILED [0.2062s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4779643Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4779946Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4780096Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4780204Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4780310Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4780507Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4780511Z 2025-12-04T09:54:35.4780591Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4783727Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4783735Z 2025-12-04T09:54:35.4783834Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4784084Z FAILED [0.2108s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4784220Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4784523Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4784699Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4784823Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4784931Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4785124Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4785127Z 2025-12-04T09:54:35.4785209Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4785373Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4785377Z 2025-12-04T09:54:35.4785473Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4785719Z FAILED [0.7343s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4785862Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4786164Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4786311Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4786433Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4786540Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4786771Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4786773Z 2025-12-04T09:54:35.4786853Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4787018Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4787020Z 2025-12-04T09:54:35.4787113Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4787362Z FAILED [0.2111s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4787502Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4787803Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4787954Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4788062Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4788168Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4788362Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4788392Z 2025-12-04T09:54:35.4788472Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4788648Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4788651Z 2025-12-04T09:54:35.4788743Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4788989Z FAILED [0.2039s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4789119Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4789421Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4789567Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4789700Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4789805Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4789998Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4790000Z 2025-12-04T09:54:35.4790079Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4790253Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4790257Z 2025-12-04T09:54:35.4790351Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4790597Z FAILED [0.3434s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4790728Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4791027Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4791175Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4791284Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4791391Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4791592Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4791594Z 2025-12-04T09:54:35.4791674Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4791835Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4791838Z 2025-12-04T09:54:35.4791930Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4792179Z FAILED [0.2148s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4792321Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4792631Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4792777Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4792884Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4792987Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4793180Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4793184Z 2025-12-04T09:54:35.4793263Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4793448Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4793450Z 2025-12-04T09:54:35.4793543Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4793790Z FAILED [0.2199s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4793920Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4794231Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4794380Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4794487Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4794593Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4794784Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4794786Z 2025-12-04T09:54:35.4794866Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4795026Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4795031Z 2025-12-04T09:54:35.4795124Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4795370Z FAILED [0.2092s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4795499Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4795794Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4795940Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4796062Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4796165Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4796370Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4796372Z 2025-12-04T09:54:35.4796450Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4796610Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4796612Z 2025-12-04T09:54:35.4796705Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4797008Z FAILED [0.2096s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4797138Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4797459Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4797607Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4797713Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4797818Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4798022Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4798026Z 2025-12-04T09:54:35.4798107Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4798268Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4798273Z 2025-12-04T09:54:35.4798365Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4798609Z FAILED [0.2078s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4798737Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4799034Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4799182Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4799289Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4799393Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4799585Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4799588Z 2025-12-04T09:54:35.4799667Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4799829Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4799847Z 2025-12-04T09:54:35.4799939Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4800198Z FAILED [0.2045s] inductor/test_compile_subprocess.py::GPUTests::test_remove_noop_slice_scatter_cuda - AssertionError: 'def forward(self, arg0_1: "Sym(s77)", arg[738 chars]13,)' != '' 2025-12-04T09:54:35.4800329Z - def forward(self, arg0_1: "Sym(s77)", arg1_1: "Sym(s27)", arg2_1: "Sym(s53)", arg3_1: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0"): 2025-12-04T09:54:35.4800626Z - empty: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.empty.memory_format([arg0_1, arg1_1, arg2_1], dtype = torch.float32, layout = torch.strided, device = device(type='cuda', index=0), pin_memory = False); arg0_1 = arg1_1 = arg2_1 = None 2025-12-04T09:54:35.4800773Z - permute: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.permute.default(empty, [0, 1, 2]); empty = permute = None 2025-12-04T09:54:35.4800880Z - add: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(arg3_1, 1); arg3_1 = None 2025-12-04T09:54:35.4800999Z - add_13: "f32[s77, s27, s53][s27*s53, s53, 1]cuda:0" = torch.ops.aten.add.Tensor(add, 1); add = None 2025-12-04T09:54:35.4801192Z - return (add_13,) : To accept the new output, re-run test with envvar EXPECTTEST_ACCEPT=1 (we recommend staging/committing your changes before doing this) 2025-12-04T09:54:35.4801195Z 2025-12-04T09:54:35.4801274Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4801434Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_remove_noop_slice_scatter_cuda 2025-12-04T09:54:35.4801438Z 2025-12-04T09:54:35.4801529Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4801748Z FAILED [0.2549s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4801752Z 2025-12-04T09:54:35.4801804Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4801920Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4802035Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4802037Z 2025-12-04T09:54:35.4802088Z The failure occurred for item [2] 2025-12-04T09:54:35.4802090Z 2025-12-04T09:54:35.4802170Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4802337Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4802339Z 2025-12-04T09:54:35.4802430Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4802632Z FAILED [0.2417s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4802635Z 2025-12-04T09:54:35.4802683Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4802793Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4802904Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4802906Z 2025-12-04T09:54:35.4802955Z The failure occurred for item [2] 2025-12-04T09:54:35.4802957Z 2025-12-04T09:54:35.4803035Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4803199Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4803201Z 2025-12-04T09:54:35.4803293Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4803488Z FAILED [0.2463s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4803510Z 2025-12-04T09:54:35.4803558Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4803665Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4803785Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4803787Z 2025-12-04T09:54:35.4803834Z The failure occurred for item [2] 2025-12-04T09:54:35.4803837Z 2025-12-04T09:54:35.4803916Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4804077Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4804079Z 2025-12-04T09:54:35.4804171Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4804369Z FAILED [0.5162s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4804371Z 2025-12-04T09:54:35.4804419Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4804521Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4804633Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4804636Z 2025-12-04T09:54:35.4804684Z The failure occurred for item [2] 2025-12-04T09:54:35.4804686Z 2025-12-04T09:54:35.4804764Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4804929Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4804931Z 2025-12-04T09:54:35.4805022Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4805234Z FAILED [0.4963s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4805237Z 2025-12-04T09:54:35.4805284Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4805384Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4805486Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4805488Z 2025-12-04T09:54:35.4805535Z The failure occurred for item [2] 2025-12-04T09:54:35.4805539Z 2025-12-04T09:54:35.4805616Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4805778Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4805780Z 2025-12-04T09:54:35.4805870Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4806065Z FAILED [0.4858s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4806067Z 2025-12-04T09:54:35.4806114Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4806215Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4806315Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4806318Z 2025-12-04T09:54:35.4806365Z The failure occurred for item [2] 2025-12-04T09:54:35.4806367Z 2025-12-04T09:54:35.4806445Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4806606Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4806608Z 2025-12-04T09:54:35.4806701Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4806939Z FAILED [0.4913s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4806941Z 2025-12-04T09:54:35.4807011Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4807110Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4807230Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4807233Z 2025-12-04T09:54:35.4807280Z The failure occurred for item [2] 2025-12-04T09:54:35.4807283Z 2025-12-04T09:54:35.4807362Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4807527Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4807530Z 2025-12-04T09:54:35.4807621Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4807815Z FAILED [0.2437s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4807820Z 2025-12-04T09:54:35.4807866Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4807976Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4808084Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4808087Z 2025-12-04T09:54:35.4808147Z The failure occurred for item [2] 2025-12-04T09:54:35.4808149Z 2025-12-04T09:54:35.4808227Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4808389Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4808391Z 2025-12-04T09:54:35.4808483Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4808677Z FAILED [0.4832s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4808679Z 2025-12-04T09:54:35.4808738Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4808841Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4808939Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4808943Z 2025-12-04T09:54:35.4808992Z The failure occurred for item [2] 2025-12-04T09:54:35.4808995Z 2025-12-04T09:54:35.4809072Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4809234Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4809236Z 2025-12-04T09:54:35.4809328Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4809521Z FAILED [0.3663s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4809523Z 2025-12-04T09:54:35.4809571Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4809681Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4809792Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4809795Z 2025-12-04T09:54:35.4809842Z The failure occurred for item [2] 2025-12-04T09:54:35.4809844Z 2025-12-04T09:54:35.4809922Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4810083Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4810085Z 2025-12-04T09:54:35.4810176Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4810370Z FAILED [0.2373s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4810372Z 2025-12-04T09:54:35.4810420Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4810527Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4810648Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4810671Z 2025-12-04T09:54:35.4810717Z The failure occurred for item [2] 2025-12-04T09:54:35.4810721Z 2025-12-04T09:54:35.4810798Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4810960Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4810962Z 2025-12-04T09:54:35.4811052Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4811244Z FAILED [0.2370s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4811246Z 2025-12-04T09:54:35.4811292Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4811398Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4811506Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4811509Z 2025-12-04T09:54:35.4811556Z The failure occurred for item [2] 2025-12-04T09:54:35.4811569Z 2025-12-04T09:54:35.4811646Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4811807Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4811809Z 2025-12-04T09:54:35.4811898Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4812096Z FAILED [0.2459s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4812098Z 2025-12-04T09:54:35.4812143Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4812260Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4812369Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4812374Z 2025-12-04T09:54:35.4812420Z The failure occurred for item [2] 2025-12-04T09:54:35.4812423Z 2025-12-04T09:54:35.4812501Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4812659Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4812661Z 2025-12-04T09:54:35.4812748Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4812935Z FAILED [0.2449s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4812937Z 2025-12-04T09:54:35.4812982Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4813084Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4813187Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4813190Z 2025-12-04T09:54:35.4813234Z The failure occurred for item [2] 2025-12-04T09:54:35.4813235Z 2025-12-04T09:54:35.4813311Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4813465Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4813467Z 2025-12-04T09:54:35.4813555Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4813740Z FAILED [0.5045s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4813743Z 2025-12-04T09:54:35.4813787Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4813885Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4813993Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4813995Z 2025-12-04T09:54:35.4814050Z The failure occurred for item [2] 2025-12-04T09:54:35.4814052Z 2025-12-04T09:54:35.4814127Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4814285Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4814287Z 2025-12-04T09:54:35.4814375Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4814565Z FAILED [0.4974s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4814567Z 2025-12-04T09:54:35.4814611Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4814709Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4814806Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4814808Z 2025-12-04T09:54:35.4814853Z The failure occurred for item [2] 2025-12-04T09:54:35.4814856Z 2025-12-04T09:54:35.4814942Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4815099Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4815101Z 2025-12-04T09:54:35.4815188Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4815373Z FAILED [0.4904s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4815375Z 2025-12-04T09:54:35.4815419Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4815527Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4815624Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4815626Z 2025-12-04T09:54:35.4815670Z The failure occurred for item [2] 2025-12-04T09:54:35.4815675Z 2025-12-04T09:54:35.4815749Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4815904Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4815906Z 2025-12-04T09:54:35.4815994Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4816179Z FAILED [0.2382s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4816181Z 2025-12-04T09:54:35.4816226Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4816328Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4816433Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4816435Z 2025-12-04T09:54:35.4816479Z The failure occurred for item [2] 2025-12-04T09:54:35.4816482Z 2025-12-04T09:54:35.4816558Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4816715Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4816719Z 2025-12-04T09:54:35.4816864Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4817049Z FAILED [0.2397s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4817051Z 2025-12-04T09:54:35.4817096Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4817198Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4817301Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4817329Z 2025-12-04T09:54:35.4817375Z The failure occurred for item [2] 2025-12-04T09:54:35.4817389Z 2025-12-04T09:54:35.4817465Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4817620Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4817622Z 2025-12-04T09:54:35.4817709Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4817895Z FAILED [0.2435s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4817897Z 2025-12-04T09:54:35.4817941Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4818044Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4818147Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4818151Z 2025-12-04T09:54:35.4818195Z The failure occurred for item [2] 2025-12-04T09:54:35.4818197Z 2025-12-04T09:54:35.4818272Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4818438Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4818440Z 2025-12-04T09:54:35.4818528Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4818713Z FAILED [0.2340s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4818715Z 2025-12-04T09:54:35.4818761Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4818861Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4818979Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4818982Z 2025-12-04T09:54:35.4819026Z The failure occurred for item [2] 2025-12-04T09:54:35.4819028Z 2025-12-04T09:54:35.4819104Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4819258Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4819260Z 2025-12-04T09:54:35.4819347Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4819532Z FAILED [0.2425s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4819535Z 2025-12-04T09:54:35.4819579Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4819683Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4819785Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4819788Z 2025-12-04T09:54:35.4819833Z The failure occurred for item [2] 2025-12-04T09:54:35.4819835Z 2025-12-04T09:54:35.4819908Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4820064Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4820066Z 2025-12-04T09:54:35.4820152Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4820338Z FAILED [0.2539s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4820341Z 2025-12-04T09:54:35.4820385Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4820486Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4820588Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4820602Z 2025-12-04T09:54:35.4820648Z The failure occurred for item [2] 2025-12-04T09:54:35.4820651Z 2025-12-04T09:54:35.4820724Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4820889Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4820891Z 2025-12-04T09:54:35.4820977Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4821164Z FAILED [0.4777s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4821166Z 2025-12-04T09:54:35.4821211Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4821306Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4821402Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4821405Z 2025-12-04T09:54:35.4821449Z The failure occurred for item [2] 2025-12-04T09:54:35.4821451Z 2025-12-04T09:54:35.4821525Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4821691Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4821693Z 2025-12-04T09:54:35.4821781Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4821966Z FAILED [0.2332s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4821968Z 2025-12-04T09:54:35.4822013Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4822115Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4822228Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4822231Z 2025-12-04T09:54:35.4822276Z The failure occurred for item [2] 2025-12-04T09:54:35.4822279Z 2025-12-04T09:54:35.4822353Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4822506Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4822509Z 2025-12-04T09:54:35.4822594Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4822775Z FAILED [0.5025s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4822777Z 2025-12-04T09:54:35.4822819Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4822913Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4823007Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4823009Z 2025-12-04T09:54:35.4823053Z The failure occurred for item [2] 2025-12-04T09:54:35.4823055Z 2025-12-04T09:54:35.4823127Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4823281Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4823283Z 2025-12-04T09:54:35.4823368Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4823549Z FAILED [0.4702s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4823551Z 2025-12-04T09:54:35.4823594Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4823687Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4823779Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4823782Z 2025-12-04T09:54:35.4823827Z The failure occurred for item [2] 2025-12-04T09:54:35.4823847Z 2025-12-04T09:54:35.4823920Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4824072Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4824085Z 2025-12-04T09:54:35.4824170Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4824352Z FAILED [0.2575s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4824354Z 2025-12-04T09:54:35.4824396Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4824497Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4824597Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4824599Z 2025-12-04T09:54:35.4824644Z The failure occurred for item [2] 2025-12-04T09:54:35.4824647Z 2025-12-04T09:54:35.4824718Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4824870Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4824885Z 2025-12-04T09:54:35.4824971Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4825152Z FAILED [1.8739s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4825154Z 2025-12-04T09:54:35.4825197Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4825291Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4825384Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4825387Z 2025-12-04T09:54:35.4825440Z The failure occurred for item [2] 2025-12-04T09:54:35.4825445Z 2025-12-04T09:54:35.4825518Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4825669Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4825671Z 2025-12-04T09:54:35.4825758Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4825938Z FAILED [0.7982s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4825940Z 2025-12-04T09:54:35.4825984Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4826077Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4826169Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4826171Z 2025-12-04T09:54:35.4826215Z The failure occurred for item [2] 2025-12-04T09:54:35.4826216Z 2025-12-04T09:54:35.4826289Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4826440Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4826443Z 2025-12-04T09:54:35.4826530Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4826710Z FAILED [0.4931s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4826712Z 2025-12-04T09:54:35.4826799Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4826893Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4826987Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4826989Z 2025-12-04T09:54:35.4827032Z The failure occurred for item [2] 2025-12-04T09:54:35.4827036Z 2025-12-04T09:54:35.4827109Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4827278Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4827294Z 2025-12-04T09:54:35.4827380Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4827565Z FAILED [0.2449s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4827567Z 2025-12-04T09:54:35.4827610Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4827715Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4827817Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4827819Z 2025-12-04T09:54:35.4827862Z The failure occurred for item [2] 2025-12-04T09:54:35.4827864Z 2025-12-04T09:54:35.4827937Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4828089Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4828092Z 2025-12-04T09:54:35.4828189Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4828372Z FAILED [0.4918s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4828374Z 2025-12-04T09:54:35.4828418Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4828511Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4828604Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4828606Z 2025-12-04T09:54:35.4828649Z The failure occurred for item [2] 2025-12-04T09:54:35.4828651Z 2025-12-04T09:54:35.4828737Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4828888Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4828891Z 2025-12-04T09:54:35.4828976Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4829157Z FAILED [0.2234s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4829159Z 2025-12-04T09:54:35.4829203Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4829302Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4829402Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4829404Z 2025-12-04T09:54:35.4829446Z The failure occurred for item [2] 2025-12-04T09:54:35.4829448Z 2025-12-04T09:54:35.4829522Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4829674Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4829677Z 2025-12-04T09:54:35.4829763Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4829949Z FAILED [0.2517s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4829951Z 2025-12-04T09:54:35.4829993Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4830093Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4830193Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4830195Z 2025-12-04T09:54:35.4830239Z The failure occurred for item [2] 2025-12-04T09:54:35.4830241Z 2025-12-04T09:54:35.4830313Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4830477Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4830480Z 2025-12-04T09:54:35.4830582Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4830764Z FAILED [0.4700s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4830766Z 2025-12-04T09:54:35.4830809Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4830904Z Greatest absolute difference: 0.58544921875 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4830996Z Greatest relative difference: 0.57568359375 at index (0, 1) (up to 0.001 allowed) 2025-12-04T09:54:35.4830998Z 2025-12-04T09:54:35.4831042Z The failure occurred for item [2] 2025-12-04T09:54:35.4831044Z 2025-12-04T09:54:35.4831118Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4831269Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4831272Z 2025-12-04T09:54:35.4831358Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4831550Z FAILED [0.2378s] inductor/test_compile_subprocess.py::GPUTests::test_var_mean_tile_reduction_True_cuda - AssertionError: Tensor-likes are not close! 2025-12-04T09:54:35.4831552Z 2025-12-04T09:54:35.4831596Z Mismatched elements: 4 / 4 (100.0%) 2025-12-04T09:54:35.4831695Z Greatest absolute difference: 0.5851404070854187 at index (0, 3) (up to 1e-05 allowed) 2025-12-04T09:54:35.4831797Z Greatest relative difference: 0.5756681561470032 at index (0, 1) (up to 1.3e-06 allowed) 2025-12-04T09:54:35.4831799Z 2025-12-04T09:54:35.4831841Z The failure occurred for item [2] 2025-12-04T09:54:35.4831843Z 2025-12-04T09:54:35.4831929Z To execute this test, run the following from the base repo dir: 2025-12-04T09:54:35.4832082Z PYTORCH_TEST_WITH_ROCM=1 python test/inductor/test_compile_subprocess.py GPUTests.test_var_mean_tile_reduction_True_cuda 2025-12-04T09:54:35.4832085Z 2025-12-04T09:54:35.4832171Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T09:54:35.4832244Z ================== 88 failed, 62 passed in 203.59s (0:03:23) =================== 2025-12-04T09:54:35.4832246Z 2025-12-04T09:54:35.4832436Z FINISHED PRINTING LOG FILE of inductor/test_compile_subprocess 3/3 (test/test-reports/inductor.test_compile_subprocess_3.3_441e1e3480e0b5ea_.log) 2025-12-04T09:54:35.4832439Z 2025-12-04T09:54:35.4832566Z Finished inductor/test_compile_subprocess 3/3 ... [2025-12-04 09:54:35.043244][2225899.500543455], took 3.59min 2025-12-04T09:54:35.4832802Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:54:35.4832891Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:54:35.4832987Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-12-04T09:54:35.4833036Z Uploading artifacts took 0.00 seconds 2025-12-04T09:54:35.4833092Z inductor/test_compile_subprocess 3/3 failed! 2025-12-04T09:54:35.4833191Z Running inductor/test_multi_kernel 1/1 ... [2025-12-04 09:54:35.054875][2225899.512197] 2025-12-04T09:54:35.4833239Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:54:35.4833613Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_multi_kernel.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:54:35.055130] 2025-12-04T09:54:41.4366398Z 2025-12-04T09:54:41.4367619Z inductor/test_multi_kernel 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_multi_kernel_1.1_2544d749af953567_.log 2025-12-04T09:54:41.4368971Z Running 0 items in this shard: 2025-12-04T09:54:41.4369123Z 2025-12-04T09:54:41.4369368Z Finished inductor/test_multi_kernel 1/1 ... [2025-12-04 09:54:41.436316][2225905.893634243], took 0.11min 2025-12-04T09:54:41.4372217Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:54:41.4485626Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:54:41.4486354Z Running inductor/test_analysis 1/1 ... [2025-12-04 09:54:41.448500][2225905.905804621] 2025-12-04T09:54:41.4486657Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:54:41.4490027Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_analysis.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:54:41.448792] 2025-12-04T09:54:47.9763511Z 2025-12-04T09:54:47.9764405Z inductor/test_analysis 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_analysis_1.1_8dadd0cc70ca2c9a_.log 2025-12-04T09:54:47.9765028Z Running 0 items in this shard: 2025-12-04T09:54:47.9769021Z 2025-12-04T09:54:47.9769231Z Finished inductor/test_analysis 1/1 ... [2025-12-04 09:54:47.976021][2225912.433340069], took 0.11min 2025-12-04T09:54:47.9769892Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:54:47.9877568Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:54:47.9882132Z Running inductor/test_pad_mm 1/1 ... [2025-12-04 09:54:47.987862][2225912.445184691] 2025-12-04T09:54:47.9882420Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:54:47.9883432Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_pad_mm.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:54:47.988136] 2025-12-04T09:54:53.9112790Z 2025-12-04T09:54:53.9113952Z inductor/test_pad_mm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_pad_mm_1.1_f86f71b53c4fe221_.log 2025-12-04T09:54:53.9114770Z Running 0 items in this shard: 2025-12-04T09:54:53.9114989Z 2025-12-04T09:54:53.9115294Z Finished inductor/test_pad_mm 1/1 ... [2025-12-04 09:54:53.910939][2225918.368258242], took 0.10min 2025-12-04T09:54:53.9118038Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:54:53.9228794Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:54:53.9231538Z Running inductor/test_triton_syntax 1/1 ... [2025-12-04 09:54:53.923001][2225918.380322111] 2025-12-04T09:54:53.9231759Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:54:53.9233894Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_triton_syntax.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:54:53.923277] 2025-12-04T09:54:59.9659430Z 2025-12-04T09:54:59.9660734Z inductor/test_triton_syntax 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_triton_syntax_1.1_ec9f1385a375b30d_.log 2025-12-04T09:54:59.9661585Z Running 0 items in this shard: 2025-12-04T09:54:59.9661803Z 2025-12-04T09:54:59.9662148Z Finished inductor/test_triton_syntax 1/1 ... [2025-12-04 09:54:59.965617][2225924.422935814], took 0.10min 2025-12-04T09:54:59.9668081Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:54:59.9772868Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:54:59.9784089Z Running test_sparse_semi_structured 1/1 ... [2025-12-04 09:54:59.977376][2225924.434700037] 2025-12-04T09:54:59.9784276Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:54:59.9784743Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_sparse_semi_structured.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:54:59.977605] 2025-12-04T09:55:06.1374286Z 2025-12-04T09:55:06.1375642Z test_sparse_semi_structured 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_sparse_semi_structured_1.1_2581abcaafe76496_.log 2025-12-04T09:55:06.1376503Z Running 0 items in this shard: 2025-12-04T09:55:06.1376717Z 2025-12-04T09:55:06.1377242Z Finished test_sparse_semi_structured 1/1 ... [2025-12-04 09:55:06.137081][2225930.594400427], took 0.10min 2025-12-04T09:55:06.1385108Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:55:06.1485894Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:55:06.1487585Z Running inductor/test_op_completeness 1/1 ... [2025-12-04 09:55:06.148502][2225930.605825635] 2025-12-04T09:55:06.1488021Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:55:06.1492312Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_op_completeness.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:55:06.148731] 2025-12-04T09:55:08.7077652Z 2025-12-04T09:55:08.7078673Z inductor/test_op_completeness 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_op_completeness_1.1_689b1a1a52cc98f1_.log 2025-12-04T09:55:08.7079271Z Running 0 items in this shard: 2025-12-04T09:55:08.7079410Z 2025-12-04T09:55:08.7079652Z Finished inductor/test_op_completeness 1/1 ... [2025-12-04 09:55:08.707435][2225933.164753245], took 0.04min 2025-12-04T09:55:08.7084803Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:55:08.7193290Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:55:08.7193897Z Running inductor/test_subgraph_choice 1/1 ... [2025-12-04 09:55:08.719234][2225933.176558078] 2025-12-04T09:55:08.7194299Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:55:08.7197091Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_subgraph_choice.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:55:08.719453] 2025-12-04T09:55:14.7945880Z 2025-12-04T09:55:14.7947341Z inductor/test_subgraph_choice 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_subgraph_choice_1.1_3b2b37b36007db5f_.log 2025-12-04T09:55:14.7948236Z Running 0 items in this shard: 2025-12-04T09:55:14.7948448Z 2025-12-04T09:55:14.7948799Z Finished inductor/test_subgraph_choice 1/1 ... [2025-12-04 09:55:14.794310][2225939.251627429], took 0.10min 2025-12-04T09:55:14.7953109Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:55:14.8058415Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:55:14.8060643Z Running inductor/test_b2b_gemm 1/1 ... [2025-12-04 09:55:14.805898][2225939.263222355] 2025-12-04T09:55:14.8061034Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:55:14.8062615Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_b2b_gemm.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:55:14.806127] 2025-12-04T09:55:21.6747649Z 2025-12-04T09:55:21.6748204Z inductor/test_b2b_gemm 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_b2b_gemm_1.1_f36f63fdd0b77120_.log 2025-12-04T09:55:21.6748759Z Running 0 items in this shard: 2025-12-04T09:55:21.6748920Z 2025-12-04T09:55:21.6749157Z Finished inductor/test_b2b_gemm 1/1 ... [2025-12-04 09:55:21.674457][2225946.131776801], took 0.11min 2025-12-04T09:55:21.6754088Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:55:21.6859698Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:55:21.6862786Z Running inductor/test_triton_heuristics 1/1 ... [2025-12-04 09:55:21.686057][2225946.143379697] 2025-12-04T09:55:21.6863138Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:55:21.6864099Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_triton_heuristics.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:55:21.686283] 2025-12-04T09:55:27.6929880Z 2025-12-04T09:55:27.6931568Z inductor/test_triton_heuristics 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_triton_heuristics_1.1_f5b010ef52ea3298_.log 2025-12-04T09:55:27.6931948Z Running 0 items in this shard: 2025-12-04T09:55:27.6932037Z 2025-12-04T09:55:27.6932173Z Finished inductor/test_triton_heuristics 1/1 ... [2025-12-04 09:55:27.692736][2225952.150055905], took 0.10min 2025-12-04T09:55:27.6935474Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:55:27.7041325Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:55:27.7043551Z Running inductor/test_ck_backend 1/1 ... [2025-12-04 09:55:27.704254][2225952.161577462] 2025-12-04T09:55:27.7043758Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:55:27.7045758Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_ck_backend.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:55:27.704478] 2025-12-04T09:55:33.3250742Z 2025-12-04T09:55:33.3251610Z inductor/test_ck_backend 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_ck_backend_1.1_9fa4255ec0fc312b_.log 2025-12-04T09:55:33.3251930Z Running 0 items in this shard: 2025-12-04T09:55:33.3252012Z 2025-12-04T09:55:33.3252133Z Finished inductor/test_ck_backend 1/1 ... [2025-12-04 09:55:33.324760][2225957.782079715], took 0.09min 2025-12-04T09:55:33.3256110Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:55:33.3365530Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:55:33.3365772Z Running test_testing 1/1 ... [2025-12-04 09:55:33.336416][2225957.79373911] 2025-12-04T09:55:33.3365954Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:55:33.3368436Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_testing.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:55:33.336658] 2025-12-04T09:55:39.4878670Z 2025-12-04T09:55:39.4879493Z test_testing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_testing_1.1_832cba6b964dcdf7_.log 2025-12-04T09:55:39.4894412Z Running 100 items in this shard: test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestTestingCUDA::test_cuda_assert_should_not_stop_common_distributed_test_suite_cuda, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies, test/test_testing.py::TestImports::test_circular_dependencies 2025-12-04T09:55:39.4906155Z 2025-12-04T09:55:39.4906255Z Finished test_testing 1/1 ... [2025-12-04 09:55:39.487653][2225963.944972651], took 0.10min 2025-12-04T09:55:39.4906638Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:55:39.4995841Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:55:39.4996558Z Running inductor/test_autoheuristic 1/1 ... [2025-12-04 09:55:39.499535][2225963.956856743] 2025-12-04T09:55:39.4996798Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:55:39.4999175Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_autoheuristic.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:55:39.499793] 2025-12-04T09:55:47.4502961Z 2025-12-04T09:55:47.4508440Z inductor/test_autoheuristic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_autoheuristic_1.1_5709f5ac6baac716_.log 2025-12-04T09:55:47.4508792Z Running 0 items in this shard: 2025-12-04T09:55:47.4508880Z 2025-12-04T09:55:47.4509013Z Finished inductor/test_autoheuristic 1/1 ... [2025-12-04 09:55:47.450063][2225971.907381448], took 0.13min 2025-12-04T09:55:47.4510009Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:55:47.4618268Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:55:47.4620240Z Running export/test_strict_export_v2 1/1 ... [2025-12-04 09:55:47.461961][2225971.919283879] 2025-12-04T09:55:47.4620444Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:55:47.4622965Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_strict_export_v2.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:55:47.462202] 2025-12-04T09:55:54.7066731Z 2025-12-04T09:55:54.7067902Z export/test_strict_export_v2 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_strict_export_v2_1.1_b927eb9a3593c315_.log 2025-12-04T09:55:54.7068526Z Running 0 items in this shard: 2025-12-04T09:55:54.7068682Z 2025-12-04T09:55:54.7068921Z Finished export/test_strict_export_v2 1/1 ... [2025-12-04 09:55:54.706405][2225979.163723009], took 0.12min 2025-12-04T09:55:54.7077129Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:55:54.7187885Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:55:54.7188664Z Running inductor/test_segmented_tree 1/1 ... [2025-12-04 09:55:54.718718][2225979.176040475] 2025-12-04T09:55:54.7192486Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:55:54.7193261Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_segmented_tree.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:55:54.718938] 2025-12-04T09:55:57.5070123Z 2025-12-04T09:55:57.5071285Z inductor/test_segmented_tree 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_segmented_tree_1.1_5cec12533bc13d02_.log 2025-12-04T09:55:57.5072179Z Running 0 items in this shard: 2025-12-04T09:55:57.5072408Z 2025-12-04T09:55:57.5072752Z Finished inductor/test_segmented_tree 1/1 ... [2025-12-04 09:55:57.506709][2225981.964029072], took 0.05min 2025-12-04T09:55:57.5076542Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:55:57.5183551Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:55:57.5185712Z Running inductor/test_kernel_optimization 1/1 ... [2025-12-04 09:55:57.518431][2225981.975754666] 2025-12-04T09:55:57.5189638Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:55:57.5190439Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_kernel_optimization.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:55:57.518655] 2025-12-04T09:56:05.6061721Z 2025-12-04T09:56:05.6062983Z inductor/test_kernel_optimization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_kernel_optimization_1.1_d8fbd8ba90f2041b_.log 2025-12-04T09:56:05.6064758Z Running 0 items in this shard: 2025-12-04T09:56:05.6064848Z 2025-12-04T09:56:05.6064989Z Finished inductor/test_kernel_optimization 1/1 ... [2025-12-04 09:56:05.605780][2225990.063098043], took 0.13min 2025-12-04T09:56:05.6069776Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:56:05.6177906Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:56:05.6180087Z Running inductor/test_inductor_scheduler 1/1 ... [2025-12-04 09:56:05.617844][2225990.075168242] 2025-12-04T09:56:05.6180311Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:56:05.6181669Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_inductor_scheduler.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:56:05.618065] 2025-12-04T09:56:13.0006945Z 2025-12-04T09:56:13.0008057Z inductor/test_inductor_scheduler 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_inductor_scheduler_1.1_afd186c27a34df43_.log 2025-12-04T09:56:13.0008946Z Running 0 items in this shard: 2025-12-04T09:56:13.0009154Z 2025-12-04T09:56:13.0009486Z Finished inductor/test_inductor_scheduler 1/1 ... [2025-12-04 09:56:13.000366][2225997.457684169], took 0.12min 2025-12-04T09:56:13.0016190Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:56:13.0124411Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:56:13.0127172Z Running inductor/test_remote_cache 1/1 ... [2025-12-04 09:56:13.012496][2225997.469819767] 2025-12-04T09:56:13.0127566Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:56:13.0129526Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_remote_cache.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:56:13.012777] 2025-12-04T09:56:14.8456402Z 2025-12-04T09:56:14.8457617Z inductor/test_remote_cache 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_remote_cache_1.1_b9ed7bcaea8047db_.log 2025-12-04T09:56:14.8458482Z Running 0 items in this shard: 2025-12-04T09:56:14.8458708Z 2025-12-04T09:56:14.8459045Z Finished inductor/test_remote_cache 1/1 ... [2025-12-04 09:56:14.845285][2225999.302603565], took 0.03min 2025-12-04T09:56:14.8464304Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:56:14.8573871Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:56:14.8575347Z Running inductor/test_coordinate_descent_tuner 1/1 ... [2025-12-04 09:56:14.857434][2225999.314758052] 2025-12-04T09:56:14.8575579Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:56:14.8578418Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_coordinate_descent_tuner.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:56:14.857657] 2025-12-04T09:56:21.5165249Z 2025-12-04T09:56:21.5166079Z inductor/test_coordinate_descent_tuner 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_coordinate_descent_tuner_1.1_4126859f40e69d1a_.log 2025-12-04T09:56:21.5167064Z Running 0 items in this shard: 2025-12-04T09:56:21.5167286Z 2025-12-04T09:56:21.5171302Z Finished inductor/test_coordinate_descent_tuner 1/1 ... [2025-12-04 09:56:21.516172][2226005.973490318], took 0.11min 2025-12-04T09:56:21.5174090Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:56:21.5289601Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:56:21.5292083Z Running inductor/test_inplace_padding 1/1 ... [2025-12-04 09:56:21.529081][2226005.986402405] 2025-12-04T09:56:21.5292334Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:56:21.5294381Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_inplace_padding.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:56:21.529342] 2025-12-04T09:59:36.0766446Z 2025-12-04T09:59:36.0767897Z inductor/test_inplace_padding 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_inplace_padding_1.1_3a1571a0e4ac804b_.log 2025-12-04T09:59:36.0781928Z Running 50 items in this shard: test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune, test/inductor/test_inplace_padding.py::InplacePaddingTest::test_linear_and_cel_max_autotune 2025-12-04T09:59:36.0788559Z 2025-12-04T09:59:36.0788701Z Finished inductor/test_inplace_padding 1/1 ... [2025-12-04 09:59:36.076213][2226200.53353014], took 3.24min 2025-12-04T09:59:36.0789103Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:59:36.0885563Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:59:36.0888066Z Running inductor/test_minifier_utils 1/1 ... [2025-12-04 09:59:36.088617][2226200.545940871] 2025-12-04T09:59:36.0889414Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:59:36.0890270Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_minifier_utils.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:59:36.088848] 2025-12-04T09:59:39.1580877Z 2025-12-04T09:59:39.1581917Z inductor/test_minifier_utils 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_minifier_utils_1.1_60b3ae48dfe48be9_.log 2025-12-04T09:59:39.1582522Z Running 0 items in this shard: 2025-12-04T09:59:39.1582663Z 2025-12-04T09:59:39.1582907Z Finished inductor/test_minifier_utils 1/1 ... [2025-12-04 09:59:39.157849][2226203.61517103], took 0.05min 2025-12-04T09:59:39.1595324Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:59:39.1688389Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:59:39.1690361Z Running inductor/test_debug_trace 1/1 ... [2025-12-04 09:59:39.168878][2226203.626202268] 2025-12-04T09:59:39.1690652Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:59:39.1691626Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_debug_trace.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:59:39.169058] 2025-12-04T09:59:48.5055857Z 2025-12-04T09:59:48.5056960Z inductor/test_debug_trace 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_debug_trace_1.1_456ffd2619706b1a_.log 2025-12-04T09:59:48.5057455Z Running 0 items in this shard: 2025-12-04T09:59:48.5057601Z 2025-12-04T09:59:48.5057792Z Finished inductor/test_debug_trace 1/1 ... [2025-12-04 09:59:48.505221][2226212.962539202], took 0.16min 2025-12-04T09:59:48.5064764Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:59:48.5179853Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:59:48.5182031Z Running dynamo/test_recompiles 1/1 ... [2025-12-04 09:59:48.518072][2226212.975391984] 2025-12-04T09:59:48.5182301Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:59:48.5184620Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_recompiles.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:59:48.518343] 2025-12-04T09:59:51.5227881Z 2025-12-04T09:59:51.5228492Z dynamo/test_recompiles 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_recompiles_1.1_166d1d4dcc61352c_.log 2025-12-04T09:59:51.5228797Z Running 0 items in this shard: 2025-12-04T09:59:51.5228877Z 2025-12-04T09:59:51.5229002Z Finished dynamo/test_recompiles 1/1 ... [2025-12-04 09:59:51.522488][2226215.979807375], took 0.05min 2025-12-04T09:59:51.5235851Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T09:59:51.5346558Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T09:59:51.5348786Z Running inductor/test_foreach 1/1 ... [2025-12-04 09:59:51.534746][2226215.992067913] 2025-12-04T09:59:51.5349599Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T09:59:51.5351149Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_foreach.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 09:59:51.535014] 2025-12-04T10:00:00.1981310Z 2025-12-04T10:00:00.1982119Z inductor/test_foreach 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_foreach_1.1_3c9486871f3acabf_.log 2025-12-04T10:00:00.1982441Z Running 0 items in this shard: 2025-12-04T10:00:00.1982525Z 2025-12-04T10:00:00.1982642Z Finished inductor/test_foreach 1/1 ... [2025-12-04 10:00:00.197883][2226224.655199404], took 0.14min 2025-12-04T10:00:00.1994353Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:00.2101475Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:00.2103531Z Running dynamo/test_buffers_override 1/1 ... [2025-12-04 10:00:00.210252][2226224.667576632] 2025-12-04T10:00:00.2111335Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:00.2111808Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_buffers_override.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:00.210465] 2025-12-04T10:00:02.1897207Z 2025-12-04T10:00:02.1898211Z dynamo/test_buffers_override 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_buffers_override_1.1_740978380611f310_.log 2025-12-04T10:00:02.1908181Z Running 0 items in this shard: 2025-12-04T10:00:02.1908315Z 2025-12-04T10:00:02.1908525Z Finished dynamo/test_buffers_override 1/1 ... [2025-12-04 10:00:02.189396][2226226.646715023], took 0.03min 2025-12-04T10:00:02.1909144Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:02.2018248Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:02.2020460Z Running inductor/test_inplacing_pass 1/1 ... [2025-12-04 10:00:02.201877][2226226.659200278] 2025-12-04T10:00:02.2020702Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:02.2022171Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_inplacing_pass.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:02.202114] 2025-12-04T10:00:10.1539636Z 2025-12-04T10:00:10.1540596Z inductor/test_inplacing_pass 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_inplacing_pass_1.1_a61133a395220a16_.log 2025-12-04T10:00:10.1550854Z Running 0 items in this shard: 2025-12-04T10:00:10.1550989Z 2025-12-04T10:00:10.1551199Z Finished inductor/test_inplacing_pass 1/1 ... [2025-12-04 10:00:10.153630][2226234.610948129], took 0.13min 2025-12-04T10:00:10.1551832Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:10.1659658Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:10.1661220Z Running inductor/test_aot_inductor_custom_ops 1/1 ... [2025-12-04 10:00:10.166026][2226234.623350039] 2025-12-04T10:00:10.1661483Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:10.1663521Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_aot_inductor_custom_ops.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:10.166248] 2025-12-04T10:00:19.6015623Z 2025-12-04T10:00:19.6016520Z inductor/test_aot_inductor_custom_ops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_aot_inductor_custom_ops_1.1_6e49d31b8b58697a_.log 2025-12-04T10:00:19.6016988Z Running 0 items in this shard: 2025-12-04T10:00:19.6017068Z 2025-12-04T10:00:19.6017208Z Finished inductor/test_aot_inductor_custom_ops 1/1 ... [2025-12-04 10:00:19.601245][2226244.058564456], took 0.16min 2025-12-04T10:00:19.6023403Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:19.6130733Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:19.6132437Z Running inductor/test_split_cat_fx_passes 1/1 ... [2025-12-04 10:00:19.613152][2226244.070475301] 2025-12-04T10:00:19.6132641Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:19.6136978Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_split_cat_fx_passes.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:19.613369] 2025-12-04T10:00:27.6836951Z 2025-12-04T10:00:27.6837914Z inductor/test_split_cat_fx_passes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_split_cat_fx_passes_1.1_7e16e82bbfcd2a32_.log 2025-12-04T10:00:27.6838540Z Running 0 items in this shard: 2025-12-04T10:00:27.6838684Z 2025-12-04T10:00:27.6844362Z Finished inductor/test_split_cat_fx_passes 1/1 ... [2025-12-04 10:00:27.683407][2226252.140726853], took 0.13min 2025-12-04T10:00:27.6845174Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:27.6953520Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:27.6955557Z Running export/test_cpp_serdes 1/1 ... [2025-12-04 10:00:27.695434][2226252.152757966] 2025-12-04T10:00:27.6955843Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:27.6957752Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_cpp_serdes.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:27.695652] 2025-12-04T10:00:35.5709606Z 2025-12-04T10:00:35.5710533Z export/test_cpp_serdes 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_cpp_serdes_1.1_4e38124e2e4b0ceb_.log 2025-12-04T10:00:35.5711073Z Running 0 items in this shard: 2025-12-04T10:00:35.5711209Z 2025-12-04T10:00:35.5711418Z Finished export/test_cpp_serdes 1/1 ... [2025-12-04 10:00:35.570617][2226260.027937042], took 0.13min 2025-12-04T10:00:35.5719158Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:35.5825659Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:35.5827802Z Running dynamo/test_pgo 1/1 ... [2025-12-04 10:00:35.582669][2226260.039992266] 2025-12-04T10:00:35.5828162Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:35.5830231Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_pgo.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:35.582902] 2025-12-04T10:00:37.7554336Z 2025-12-04T10:00:37.7555118Z dynamo/test_pgo 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_pgo_1.1_fea5af0bd3644147_.log 2025-12-04T10:00:37.7555831Z Running 0 items in this shard: 2025-12-04T10:00:37.7555916Z 2025-12-04T10:00:37.7556023Z Finished dynamo/test_pgo 1/1 ... [2025-12-04 10:00:37.755185][2226262.212501917], took 0.04min 2025-12-04T10:00:37.7565201Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:37.7674737Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:37.7676818Z Running dynamo/test_config 1/1 ... [2025-12-04 10:00:37.767618][2226262.224941091] 2025-12-04T10:00:37.7677007Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:37.7679607Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_config.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:37.767862] 2025-12-04T10:00:41.2809892Z 2025-12-04T10:00:41.2811833Z dynamo/test_config 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_config_1.1_4707744602aa4337_.log 2025-12-04T10:00:41.2812629Z Running 0 items in this shard: 2025-12-04T10:00:41.2812856Z 2025-12-04T10:00:41.2813169Z Finished dynamo/test_config 1/1 ... [2025-12-04 10:00:41.280655][2226265.737972794], took 0.06min 2025-12-04T10:00:41.2819842Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:41.2939126Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:41.2940070Z Running dynamo/test_metrics_context 1/1 ... [2025-12-04 10:00:41.293933][2226265.751250519] 2025-12-04T10:00:41.2940431Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:41.2943811Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_metrics_context.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:41.294243] 2025-12-04T10:00:44.3077997Z 2025-12-04T10:00:44.3078866Z dynamo/test_metrics_context 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_metrics_context_1.1_f7d94e2ae3364fa1_.log 2025-12-04T10:00:44.3079311Z Running 0 items in this shard: 2025-12-04T10:00:44.3079413Z 2025-12-04T10:00:44.3079574Z Finished dynamo/test_metrics_context 1/1 ... [2025-12-04 10:00:44.307593][2226268.764910218], took 0.05min 2025-12-04T10:00:44.3086960Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:44.3193921Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:44.3197526Z Running export/test_package 1/1 ... [2025-12-04 10:00:44.319527][2226268.776850467] 2025-12-04T10:00:44.3197890Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:44.3198612Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_package.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:44.319756] 2025-12-04T10:00:46.2675462Z 2025-12-04T10:00:46.2676419Z export/test_package 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_package_1.1_757331b17095c9ea_.log 2025-12-04T10:00:46.2677151Z Running 0 items in this shard: 2025-12-04T10:00:46.2678522Z 2025-12-04T10:00:46.2678740Z Finished export/test_package 1/1 ... [2025-12-04 10:00:46.267297][2226270.724616228], took 0.03min 2025-12-04T10:00:46.2684135Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:46.2794523Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:46.2796977Z Running export/test_export_opinfo 1/1 ... [2025-12-04 10:00:46.279571][2226270.736892459] 2025-12-04T10:00:46.2797303Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:46.2800671Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'export/test_export_opinfo.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:46.279934] 2025-12-04T10:00:50.2750706Z 2025-12-04T10:00:50.2751869Z export/test_export_opinfo 1/1 was successful, full logs can be found in artifacts with path test/test-reports/export.test_export_opinfo_1.1_fa907d48ae6775cf_.log 2025-12-04T10:00:50.2752530Z Running 0 items in this shard: 2025-12-04T10:00:50.2752696Z 2025-12-04T10:00:50.2758263Z Finished export/test_export_opinfo 1/1 ... [2025-12-04 10:00:50.274713][2226274.732030751], took 0.07min 2025-12-04T10:00:50.2762494Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:50.2871460Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:50.2873673Z Running dynamo/test_nops 1/1 ... [2025-12-04 10:00:50.287248][2226274.744570496] 2025-12-04T10:00:50.2873961Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:50.2882876Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_nops.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:50.287491] 2025-12-04T10:00:52.4542162Z 2025-12-04T10:00:52.4543017Z dynamo/test_nops 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_nops_1.1_f979545477755402_.log 2025-12-04T10:00:52.4543309Z Running 0 items in this shard: 2025-12-04T10:00:52.4543388Z 2025-12-04T10:00:52.4543501Z Finished dynamo/test_nops 1/1 ... [2025-12-04 10:00:52.453910][2226276.911228574], took 0.04min 2025-12-04T10:00:52.4556574Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:00:52.4659348Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:00:52.4661602Z Running inductor/test_graph_transform_observer 1/1 ... [2025-12-04 10:00:52.466032][2226276.92335489] 2025-12-04T10:00:52.4661841Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:00:52.4663924Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_graph_transform_observer.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:00:52.466274] 2025-12-04T10:01:00.1520781Z 2025-12-04T10:01:00.1521605Z inductor/test_graph_transform_observer 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_graph_transform_observer_1.1_7d3a0cdc9d67ebd0_.log 2025-12-04T10:01:00.1521981Z Running 0 items in this shard: 2025-12-04T10:01:00.1522072Z 2025-12-04T10:01:00.1522563Z Finished inductor/test_graph_transform_observer 1/1 ... [2025-12-04 10:01:00.151876][2226284.609196066], took 0.13min 2025-12-04T10:01:00.1530688Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:01:00.1638311Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:01:00.1640622Z Running inductor/test_memory 1/1 ... [2025-12-04 10:01:00.163922][2226284.621246516] 2025-12-04T10:01:00.1640911Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:01:00.1642564Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_memory.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:01:00.164132] 2025-12-04T10:01:07.1863063Z 2025-12-04T10:01:07.1863896Z inductor/test_memory 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_memory_1.1_fd2c08b05c82f4b7_.log 2025-12-04T10:01:07.1864218Z Running 0 items in this shard: 2025-12-04T10:01:07.1864312Z 2025-12-04T10:01:07.1864437Z Finished inductor/test_memory 1/1 ... [2025-12-04 10:01:07.186048][2226291.643367396], took 0.12min 2025-12-04T10:01:07.1878485Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:01:07.1981305Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:01:07.1984335Z Running inductor/test_control_flow 1/4 ... [2025-12-04 10:01:07.198267][2226291.655589303] 2025-12-04T10:01:07.1984545Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:01:07.1986343Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_control_flow.py', '--shard-id=1', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:01:07.198510] 2025-12-04T10:01:14.9909620Z 2025-12-04T10:01:14.9911358Z inductor/test_control_flow 1/4 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_control_flow_1.4_dabc57682fc3ded7_.log 2025-12-04T10:01:14.9912277Z Running 0 items in this shard: 2025-12-04T10:01:14.9912506Z 2025-12-04T10:01:14.9912849Z Finished inductor/test_control_flow 1/4 ... [2025-12-04 10:01:14.990572][2226299.447891705], took 0.13min 2025-12-04T10:01:14.9917369Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:01:15.0025552Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:01:15.0027384Z Running inductor/test_pallas 1/1 ... [2025-12-04 10:01:15.002548][2226299.459872199] 2025-12-04T10:01:15.0027728Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:01:15.0029353Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'inductor/test_pallas.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:01:15.002783] 2025-12-04T10:01:23.4014231Z 2025-12-04T10:01:23.4015518Z inductor/test_pallas 1/1 was successful, full logs can be found in artifacts with path test/test-reports/inductor.test_pallas_1.1_1a4d78abece1f7e4_.log 2025-12-04T10:01:23.4016205Z 2025-12-04T10:01:23.4016539Z Finished inductor/test_pallas 1/1 ... [2025-12-04 10:01:23.401093][2226307.858411366], took 0.14min 2025-12-04T10:01:23.4025189Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:01:23.4139441Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:01:23.4142110Z Running dynamo/test_callback 1/1 ... [2025-12-04 10:01:23.414086][2226307.871407927] 2025-12-04T10:01:23.4142975Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:01:23.4144649Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_callback.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:01:23.414352] 2025-12-04T10:01:31.1879088Z 2025-12-04T10:01:31.1879954Z dynamo/test_callback 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_callback_1.1_0d5a112d877c421f_.log 2025-12-04T10:01:31.1880383Z Running 0 items in this shard: 2025-12-04T10:01:31.1880505Z 2025-12-04T10:01:31.1880661Z Finished dynamo/test_callback 1/1 ... [2025-12-04 10:01:31.187541][2226315.644858738], took 0.13min 2025-12-04T10:01:31.1890903Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:01:31.2000142Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:01:31.2002414Z Running dynamo/test_cudagraphs_expandable_segments 1/1 ... [2025-12-04 10:01:31.200108][2226315.657430071] 2025-12-04T10:01:31.2004391Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:01:31.2005011Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_cudagraphs_expandable_segments.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:01:31.200364] 2025-12-04T10:01:33.5570656Z 2025-12-04T10:01:33.5572049Z dynamo/test_cudagraphs_expandable_segments 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_cudagraphs_expandable_segments_1.1_7db242d57afb96c5_.log 2025-12-04T10:01:33.5573725Z Running 0 items in this shard: 2025-12-04T10:01:33.5573963Z 2025-12-04T10:01:33.5574374Z Finished dynamo/test_cudagraphs_expandable_segments 1/1 ... [2025-12-04 10:01:33.556760][2226318.014077745], took 0.04min 2025-12-04T10:01:33.5583366Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:01:33.5692787Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:01:33.5694969Z Running dynamo/test_aot_autograd_cache 1/1 ... [2025-12-04 10:01:33.569316][2226318.026639649] 2025-12-04T10:01:33.5695317Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:01:33.5697098Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'dynamo/test_aot_autograd_cache.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:01:33.569540] 2025-12-04T10:01:39.6459565Z 2025-12-04T10:01:39.6460719Z dynamo/test_aot_autograd_cache 1/1 was successful, full logs can be found in artifacts with path test/test-reports/dynamo.test_aot_autograd_cache_1.1_605d0d5b5c74b7c0_.log 2025-12-04T10:01:39.6461676Z Running 0 items in this shard: 2025-12-04T10:01:39.6461903Z 2025-12-04T10:01:39.6462250Z Finished dynamo/test_aot_autograd_cache 1/1 ... [2025-12-04 10:01:39.645636][2226324.102952616], took 0.10min 2025-12-04T10:01:39.6475058Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:01:39.6585549Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:01:39.6587621Z Running test_utils_config_module 1/1 ... [2025-12-04 10:01:39.658609][2226324.115932511] 2025-12-04T10:01:39.6587961Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:01:39.6589821Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_utils_config_module.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:01:39.658843] 2025-12-04T10:01:41.7783199Z 2025-12-04T10:01:41.7784122Z test_utils_config_module 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_utils_config_module_1.1_404b0f99bdfcdada_.log 2025-12-04T10:01:41.7784725Z Running 0 items in this shard: 2025-12-04T10:01:41.7784889Z 2025-12-04T10:01:41.7785114Z Finished test_utils_config_module 1/1 ... [2025-12-04 10:01:41.777997][2226326.235315159], took 0.04min 2025-12-04T10:01:41.7799009Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:01:41.7908772Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:01:41.7910279Z Running test_hop_infra 1/1 ... [2025-12-04 10:01:41.790924][2226326.248248416] 2025-12-04T10:01:41.7910462Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:01:41.7913757Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_hop_infra.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:01:41.791145] 2025-12-04T10:01:44.4187845Z 2025-12-04T10:01:44.4188753Z test_hop_infra 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_hop_infra_1.1_affed15be16b21c3_.log 2025-12-04T10:01:44.4189329Z Running 0 items in this shard: 2025-12-04T10:01:44.4189487Z 2025-12-04T10:01:44.4189697Z Finished test_hop_infra 1/1 ... [2025-12-04 10:01:44.418469][2226328.875788898], took 0.04min 2025-12-04T10:01:44.4202376Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:01:44.4307935Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:01:44.4310568Z Running test_transformers 1/1 ... [2025-12-04 10:01:44.430727][2226328.888051282] 2025-12-04T10:01:44.4310893Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:01:44.4311681Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_transformers.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:01:44.430939] 2025-12-04T10:01:59.7434323Z 2025-12-04T10:01:59.7435186Z test_transformers 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_transformers_1.1_b07e97c9ea20ff07_.log 2025-12-04T10:01:59.7435903Z Running 0 items in this shard: 2025-12-04T10:01:59.7436074Z 2025-12-04T10:01:59.7436315Z Finished test_transformers 1/1 ... [2025-12-04 10:01:59.743153][2226344.20047225], took 0.26min 2025-12-04T10:01:59.7446206Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:01:59.7554682Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:01:59.7556802Z Running torch_np/numpy_tests/core/test_scalarinherit 1/1 ... [2025-12-04 10:01:59.755523][2226344.212846174] 2025-12-04T10:01:59.7557218Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:01:59.7559076Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_scalarinherit.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:01:59.755756] 2025-12-04T10:02:02.4779570Z 2025-12-04T10:02:02.4782144Z torch_np/numpy_tests/core/test_scalarinherit 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_scalarinherit_1.1_c7091fa3f46b7848_.log 2025-12-04T10:02:02.4783277Z Running 0 items in this shard: 2025-12-04T10:02:02.4783508Z 2025-12-04T10:02:02.4783921Z Finished torch_np/numpy_tests/core/test_scalarinherit 1/1 ... [2025-12-04 10:02:02.477622][2226346.934941182], took 0.05min 2025-12-04T10:02:02.4794418Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:02:02.4905145Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:02:02.4907182Z Running test_tensorexpr_pybind 1/1 ... [2025-12-04 10:02:02.490549][2226346.947873183] 2025-12-04T10:02:02.4907556Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:02:02.4909302Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_tensorexpr_pybind.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:02:02.490775] 2025-12-04T10:02:04.6263555Z 2025-12-04T10:02:04.6264179Z test_tensorexpr_pybind 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_tensorexpr_pybind_1.1_4e0c0ce3181c7e29_.log 2025-12-04T10:02:04.6264580Z Running 0 items in this shard: 2025-12-04T10:02:04.6264684Z 2025-12-04T10:02:04.6264821Z Finished test_tensorexpr_pybind 1/1 ... [2025-12-04 10:02:04.625980][2226349.08330032], took 0.04min 2025-12-04T10:02:04.6273152Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:02:04.6382975Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:02:04.6384715Z Running test_fx_experimental 1/1 ... [2025-12-04 10:02:04.638392][2226349.095715584] 2025-12-04T10:02:04.6384948Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:02:04.6387313Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_fx_experimental.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:02:04.638617] 2025-12-04T10:02:09.1015884Z 2025-12-04T10:02:09.1016974Z test_fx_experimental 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_fx_experimental_1.1_fc17588850b550f8_.log 2025-12-04T10:02:09.1017588Z Running 0 items in this shard: 2025-12-04T10:02:09.1017752Z 2025-12-04T10:02:09.1017989Z Finished test_fx_experimental 1/1 ... [2025-12-04 10:02:09.101273][2226353.558590921], took 0.07min 2025-12-04T10:02:09.1029717Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:02:09.1140716Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:02:09.1142609Z Running test_namedtensor 1/1 ... [2025-12-04 10:02:09.114106][2226353.571429686] 2025-12-04T10:02:09.1142945Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:02:09.1144708Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_namedtensor.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:02:09.114335] 2025-12-04T10:02:11.5341075Z 2025-12-04T10:02:11.5341709Z test_namedtensor 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_namedtensor_1.1_01536deb53cb92e3_.log 2025-12-04T10:02:11.5342468Z Running 0 items in this shard: 2025-12-04T10:02:11.5343051Z 2025-12-04T10:02:11.5343361Z Finished test_namedtensor 1/1 ... [2025-12-04 10:02:11.533805][2226355.991124381], took 0.04min 2025-12-04T10:02:11.5355145Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:02:11.5465226Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:02:11.5466943Z Running test_dataloader 1/1 ... [2025-12-04 10:02:11.546553][2226356.003876628] 2025-12-04T10:02:11.5467284Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:02:11.5469290Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_dataloader.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:02:11.546775] 2025-12-04T10:02:14.8424142Z 2025-12-04T10:02:14.8425164Z test_dataloader 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_dataloader_1.1_5d880b35f781b99b_.log 2025-12-04T10:02:14.8425939Z Running 0 items in this shard: 2025-12-04T10:02:14.8426166Z 2025-12-04T10:02:14.8427249Z Finished test_dataloader 1/1 ... [2025-12-04 10:02:14.842146][2226359.299464196], took 0.05min 2025-12-04T10:02:14.8438970Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:02:14.8549426Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:02:14.8551199Z Running test_decomp 6/12 ... [2025-12-04 10:02:14.854987][2226359.312310962] 2025-12-04T10:02:14.8551480Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:02:14.8553627Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '--shard-id=6', '--num-shards=12', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:02:14.855220] 2025-12-04T10:02:27.2950387Z 2025-12-04T10:02:27.2951456Z test_decomp 6/12 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_6.12_806f6d5ab86932ae_.log 2025-12-04T10:02:27.2959254Z Running 0 items in this shard: 2025-12-04T10:02:27.2959524Z 2025-12-04T10:02:27.2959748Z Finished test_decomp 6/12 ... [2025-12-04 10:02:27.294710][2226371.752027657], took 0.21min 2025-12-04T10:02:27.2960839Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:02:27.3075219Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:02:27.3076672Z Running test_decomp 12/12 ... [2025-12-04 10:02:27.307502][2226371.764824227] 2025-12-04T10:02:27.3077040Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:02:27.3078992Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_decomp.py', '--shard-id=12', '--num-shards=12', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:02:27.307762] 2025-12-04T10:06:48.9675437Z 2025-12-04T10:06:48.9676251Z test_decomp 12/12 was successful, full logs can be found in artifacts with path test/test-reports/test_decomp_12.12_5f1c890e505cfa19_.log 2025-12-04T10:06:48.9687364Z Running 50 items in this shard: test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64, test/test_decomp.py::TestDecompCUDA::test_comprehensive_nn_functional_nll_loss_cuda_float64 2025-12-04T10:06:48.9695775Z 2025-12-04T10:06:48.9695914Z Finished test_decomp 12/12 ... [2025-12-04 10:06:48.966913][2226633.424229982], took 4.36min 2025-12-04T10:06:48.9696358Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:06:48.9796246Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:06:48.9798375Z Running test_ci_sanity_check_fail 1/1 ... [2025-12-04 10:06:48.979670][2226633.436994411] 2025-12-04T10:06:48.9798586Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:06:48.9800010Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ci_sanity_check_fail.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:06:48.979898] 2025-12-04T10:06:50.9478270Z Finished test_ci_sanity_check_fail 1/1 ... [2025-12-04 10:06:50.947325][2226635.404626773], took 0.03min 2025-12-04T10:06:50.9489514Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:06:50.9598866Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:06:50.9599231Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-12-04T10:06:50.9599530Z Uploading artifacts took 0.00 seconds 2025-12-04T10:06:50.9599739Z test_ci_sanity_check_fail 1/1 failed! 2025-12-04T10:06:50.9602909Z Running test_meta 3/3 ... [2025-12-04 10:06:50.959983][2226635.417307434] 2025-12-04T10:06:50.9603183Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:06:50.9605962Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_meta.py', '--shard-id=3', '--num-shards=3', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:06:50.960218] 2025-12-04T10:07:55.1264178Z 2025-12-04T10:07:55.1264895Z test_meta 3/3 was successful, full logs can be found in artifacts with path test/test-reports/test_meta_3.3_f9534c273fa76bd7_.log 2025-12-04T10:07:55.1265229Z Running 0 items in this shard: 2025-12-04T10:07:55.1265330Z 2025-12-04T10:07:55.1265437Z Finished test_meta 3/3 ... [2025-12-04 10:07:55.126131][2226699.583450823], took 1.07min 2025-12-04T10:07:55.1274967Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:07:55.1396676Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:07:55.1399302Z Running test_ops_gradients 2/2 ... [2025-12-04 10:07:55.139855][2226699.597177171] 2025-12-04T10:07:55.1399497Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:07:55.1402095Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops_gradients.py', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:07:55.140138] 2025-12-04T10:08:04.5569631Z 2025-12-04T10:08:04.5570544Z test_ops_gradients 2/2 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_gradients_2.2_4219eceb223ba915_.log 2025-12-04T10:08:04.5570995Z Running 0 items in this shard: 2025-12-04T10:08:04.5571117Z 2025-12-04T10:08:04.5571283Z Finished test_ops_gradients 2/2 ... [2025-12-04 10:08:04.556726][2226709.014046302], took 0.16min 2025-12-04T10:08:04.5580844Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:08:04.5691133Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:08:04.5693735Z Running test_linalg 2/2 ... [2025-12-04 10:08:04.569221][2226709.026544625] 2025-12-04T10:08:04.5693917Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:08:04.5695557Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_linalg.py', '--shard-id=2', '--num-shards=2', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:08:04.569459] 2025-12-04T10:08:48.0210129Z 2025-12-04T10:08:48.0210967Z PRINTING LOG FILE of test_linalg 2/2 (test/test-reports/test_linalg_2.2_4f98d9a906e7bcd5_.log) 2025-12-04T10:08:48.0211405Z Test results will be stored in test-reports/python-pytest/test_linalg/test_linalg-115bbfbfd2646928.xml 2025-12-04T10:08:48.0211825Z ============================= test session starts ============================== 2025-12-04T10:08:48.0212153Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:08:48.0212424Z cachedir: .pytest_cache 2025-12-04T10:08:48.0227225Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:08:48.0227500Z rootdir: /var/lib/jenkins/pytorch 2025-12-04T10:08:48.0227638Z configfile: pytest.ini 2025-12-04T10:08:48.0227898Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:08:48.0228171Z collecting ... collected 1263 items 2025-12-04T10:08:48.0228333Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T10:08:48.0235055Z Running 50 items in this shard: test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32, test/test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0241452Z 2025-12-04T10:08:48.0241594Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 PASSED [37.8697s] [ 2%] 2025-12-04T10:08:48.0241906Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0448s] [ 2%] 2025-12-04T10:08:48.0242212Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0497s] [ 2%] 2025-12-04T10:08:48.0242516Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0025s] [ 2%] 2025-12-04T10:08:48.0242817Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0021s] [ 2%] 2025-12-04T10:08:48.0243126Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0022s] [ 2%] 2025-12-04T10:08:48.0243425Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0020s] [ 2%] 2025-12-04T10:08:48.0243726Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0021s] [ 2%] 2025-12-04T10:08:48.0244028Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0018s] [ 2%] 2025-12-04T10:08:48.0244328Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0020s] [ 2%] 2025-12-04T10:08:48.0244629Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0018s] [ 2%] 2025-12-04T10:08:48.0244957Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0020s] [ 2%] 2025-12-04T10:08:48.0245258Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0018s] [ 2%] 2025-12-04T10:08:48.0245585Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0020s] [ 2%] 2025-12-04T10:08:48.0245886Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0018s] [ 2%] 2025-12-04T10:08:48.0246186Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0019s] [ 2%] 2025-12-04T10:08:48.0246484Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0017s] [ 2%] 2025-12-04T10:08:48.0246825Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0019s] [ 2%] 2025-12-04T10:08:48.0247128Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0017s] [ 2%] 2025-12-04T10:08:48.0247433Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0019s] [ 2%] 2025-12-04T10:08:48.0247755Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0018s] [ 2%] 2025-12-04T10:08:48.0248057Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0019s] [ 2%] 2025-12-04T10:08:48.0248355Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0017s] [ 2%] 2025-12-04T10:08:48.0248661Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0019s] [ 2%] 2025-12-04T10:08:48.0248960Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0017s] [ 2%] 2025-12-04T10:08:48.0249274Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0020s] [ 2%] 2025-12-04T10:08:48.0249574Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0018s] [ 2%] 2025-12-04T10:08:48.0249877Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0019s] [ 2%] 2025-12-04T10:08:48.0250174Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0017s] [ 2%] 2025-12-04T10:08:48.0250471Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0019s] [ 2%] 2025-12-04T10:08:48.0250770Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0017s] [ 2%] 2025-12-04T10:08:48.0251069Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0021s] [ 2%] 2025-12-04T10:08:48.0251371Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0010s] [ 2%] 2025-12-04T10:08:48.0251672Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0251972Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0252273Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0252574Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0008s] [ 2%] 2025-12-04T10:08:48.0252873Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0253175Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0253475Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0008s] [ 2%] 2025-12-04T10:08:48.0253797Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0254097Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0254415Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0008s] [ 2%] 2025-12-04T10:08:48.0254729Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0255022Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0255314Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0255607Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0255898Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0007s] [ 2%] 2025-12-04T10:08:48.0256193Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0009s] [ 2%] 2025-12-04T10:08:48.0256498Z test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 FAILED [0.0008s] [ 2%] 2025-12-04T10:08:48.0256665Z 2025-12-04T10:08:48.0256724Z =================================== FAILURES =================================== 2025-12-04T10:08:48.0256947Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0257125Z Traceback (most recent call last): 2025-12-04T10:08:48.0257347Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0257577Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0257809Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0258022Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0258214Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0258393Z raise self.failureException(msg) 2025-12-04T10:08:48.0258523Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0258602Z 2025-12-04T10:08:48.0258681Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0258945Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0259133Z 2025-12-04T10:08:48.0259224Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0259442Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0259620Z Traceback (most recent call last): 2025-12-04T10:08:48.0259832Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0260059Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0260260Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0260463Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0260648Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0260823Z raise self.failureException(msg) 2025-12-04T10:08:48.0260946Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0261023Z 2025-12-04T10:08:48.0261098Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0261358Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0261544Z 2025-12-04T10:08:48.0261634Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0261849Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0262070Z Traceback (most recent call last): 2025-12-04T10:08:48.0262280Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0262523Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0262721Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0262919Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0263102Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0263275Z raise self.failureException(msg) 2025-12-04T10:08:48.0263398Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0263475Z 2025-12-04T10:08:48.0263551Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0263809Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0263997Z 2025-12-04T10:08:48.0264088Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0264304Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0264498Z Traceback (most recent call last): 2025-12-04T10:08:48.0264707Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0264931Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0265130Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0265327Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0265511Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0265682Z raise self.failureException(msg) 2025-12-04T10:08:48.0265830Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0265912Z 2025-12-04T10:08:48.0265986Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0266245Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0266435Z 2025-12-04T10:08:48.0266520Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0266734Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0266946Z Traceback (most recent call last): 2025-12-04T10:08:48.0267156Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0267378Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0267583Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0267782Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0267968Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0268141Z raise self.failureException(msg) 2025-12-04T10:08:48.0268266Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0268344Z 2025-12-04T10:08:48.0268419Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0268677Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0268865Z 2025-12-04T10:08:48.0268950Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0269166Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0269341Z Traceback (most recent call last): 2025-12-04T10:08:48.0269553Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0269777Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0269993Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0270192Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0270394Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0270567Z raise self.failureException(msg) 2025-12-04T10:08:48.0270690Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0270769Z 2025-12-04T10:08:48.0270842Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0271100Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0271287Z 2025-12-04T10:08:48.0271372Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0271587Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0271763Z Traceback (most recent call last): 2025-12-04T10:08:48.0271970Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0272195Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0272408Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0272607Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0272790Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0272963Z raise self.failureException(msg) 2025-12-04T10:08:48.0273086Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0273163Z 2025-12-04T10:08:48.0273237Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0273514Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0273702Z 2025-12-04T10:08:48.0273789Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0274002Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0274178Z Traceback (most recent call last): 2025-12-04T10:08:48.0274388Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0274613Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0274813Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0275012Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0275195Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0275370Z raise self.failureException(msg) 2025-12-04T10:08:48.0275493Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0275572Z 2025-12-04T10:08:48.0275647Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0275907Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0276094Z 2025-12-04T10:08:48.0276184Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0276399Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0276573Z Traceback (most recent call last): 2025-12-04T10:08:48.0276814Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0277037Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0277234Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0277431Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0277615Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0277815Z raise self.failureException(msg) 2025-12-04T10:08:48.0277938Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0278029Z 2025-12-04T10:08:48.0278103Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0278361Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0278546Z 2025-12-04T10:08:48.0278632Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0278846Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0279020Z Traceback (most recent call last): 2025-12-04T10:08:48.0279228Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0279453Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0279653Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0279856Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0280040Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0280232Z raise self.failureException(msg) 2025-12-04T10:08:48.0280356Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0280434Z 2025-12-04T10:08:48.0280507Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0280765Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0280954Z 2025-12-04T10:08:48.0281042Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0281256Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0281447Z Traceback (most recent call last): 2025-12-04T10:08:48.0281656Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0281884Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0282085Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0282287Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0282473Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0282647Z raise self.failureException(msg) 2025-12-04T10:08:48.0282770Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0282849Z 2025-12-04T10:08:48.0282921Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0283179Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0283371Z 2025-12-04T10:08:48.0283459Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0283676Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0283852Z Traceback (most recent call last): 2025-12-04T10:08:48.0284063Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0284289Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0284488Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0284688Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0284872Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0285048Z raise self.failureException(msg) 2025-12-04T10:08:48.0285169Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0285248Z 2025-12-04T10:08:48.0285322Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0285580Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0285787Z 2025-12-04T10:08:48.0285873Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0286107Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0286284Z Traceback (most recent call last): 2025-12-04T10:08:48.0286498Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0286723Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0286950Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0287148Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0287332Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0287507Z raise self.failureException(msg) 2025-12-04T10:08:48.0287632Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0287710Z 2025-12-04T10:08:48.0287782Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0288063Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0288252Z 2025-12-04T10:08:48.0288338Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0288553Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0288729Z Traceback (most recent call last): 2025-12-04T10:08:48.0288937Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0289159Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0289380Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0289582Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0289765Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0289944Z raise self.failureException(msg) 2025-12-04T10:08:48.0290068Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0290147Z 2025-12-04T10:08:48.0290223Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0290553Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0290757Z 2025-12-04T10:08:48.0290870Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0291115Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0291516Z Traceback (most recent call last): 2025-12-04T10:08:48.0291771Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0292047Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0292301Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0292538Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0292784Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0293001Z raise self.failureException(msg) 2025-12-04T10:08:48.0293157Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0293285Z 2025-12-04T10:08:48.0293371Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0293670Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0293887Z 2025-12-04T10:08:48.0293999Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0294261Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0301231Z Traceback (most recent call last): 2025-12-04T10:08:48.0301458Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0301734Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0301934Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0302140Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0302322Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0302507Z raise self.failureException(msg) 2025-12-04T10:08:48.0302629Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0302706Z 2025-12-04T10:08:48.0302782Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0303043Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0303233Z 2025-12-04T10:08:48.0303322Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0303538Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0303763Z Traceback (most recent call last): 2025-12-04T10:08:48.0303979Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0304208Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0304410Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0304610Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0304794Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0304972Z raise self.failureException(msg) 2025-12-04T10:08:48.0305118Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0305200Z 2025-12-04T10:08:48.0305277Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0305541Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0305731Z 2025-12-04T10:08:48.0305824Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0306043Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0306221Z Traceback (most recent call last): 2025-12-04T10:08:48.0306438Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0306663Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0306911Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0307116Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0307305Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0307485Z raise self.failureException(msg) 2025-12-04T10:08:48.0307611Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0307691Z 2025-12-04T10:08:48.0307771Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0308034Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0308223Z 2025-12-04T10:08:48.0308313Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0308533Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0308710Z Traceback (most recent call last): 2025-12-04T10:08:48.0308922Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0309152Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0309378Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0309577Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0309783Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0309962Z raise self.failureException(msg) 2025-12-04T10:08:48.0310090Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0310170Z 2025-12-04T10:08:48.0310245Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0310509Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0310700Z 2025-12-04T10:08:48.0310789Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0311008Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0311189Z Traceback (most recent call last): 2025-12-04T10:08:48.0311409Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0311639Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0311861Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0312065Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0312252Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0312430Z raise self.failureException(msg) 2025-12-04T10:08:48.0312558Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0312639Z 2025-12-04T10:08:48.0312714Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0312976Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0313179Z 2025-12-04T10:08:48.0313266Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0313487Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0313668Z Traceback (most recent call last): 2025-12-04T10:08:48.0313883Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0314113Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0314314Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0314516Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0314701Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0314879Z raise self.failureException(msg) 2025-12-04T10:08:48.0315007Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0315088Z 2025-12-04T10:08:48.0315163Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0315425Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0315615Z 2025-12-04T10:08:48.0315702Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0315922Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0316100Z Traceback (most recent call last): 2025-12-04T10:08:48.0316311Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0316540Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0316768Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0316969Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0317157Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0317362Z raise self.failureException(msg) 2025-12-04T10:08:48.0317487Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0317565Z 2025-12-04T10:08:48.0317658Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0317923Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0318110Z 2025-12-04T10:08:48.0318202Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0318420Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0318598Z Traceback (most recent call last): 2025-12-04T10:08:48.0318815Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0319043Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0319246Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0319451Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0319638Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0319818Z raise self.failureException(msg) 2025-12-04T10:08:48.0319963Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0320042Z 2025-12-04T10:08:48.0320121Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0320384Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0320570Z 2025-12-04T10:08:48.0320661Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0320879Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0321057Z Traceback (most recent call last): 2025-12-04T10:08:48.0321287Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0321519Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0321720Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0321926Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0322114Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0322293Z raise self.failureException(msg) 2025-12-04T10:08:48.0322419Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0322497Z 2025-12-04T10:08:48.0322575Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0322838Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0323025Z 2025-12-04T10:08:48.0323118Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0323340Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0323518Z Traceback (most recent call last): 2025-12-04T10:08:48.0323733Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0323962Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0324165Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0324367Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0324554Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0324728Z raise self.failureException(msg) 2025-12-04T10:08:48.0324856Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0324937Z 2025-12-04T10:08:48.0325011Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0325277Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0331537Z 2025-12-04T10:08:48.0331624Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0331870Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0332052Z Traceback (most recent call last): 2025-12-04T10:08:48.0332266Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0332495Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0332699Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0332905Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0333094Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0333280Z raise self.failureException(msg) 2025-12-04T10:08:48.0333417Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0333507Z 2025-12-04T10:08:48.0333584Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0333874Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0334071Z 2025-12-04T10:08:48.0334160Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0334387Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0334573Z Traceback (most recent call last): 2025-12-04T10:08:48.0334794Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0335029Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0335256Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0335467Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0335665Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0335850Z raise self.failureException(msg) 2025-12-04T10:08:48.0335987Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0336075Z 2025-12-04T10:08:48.0336153Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0336424Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0336621Z 2025-12-04T10:08:48.0336711Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0336976Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0337162Z Traceback (most recent call last): 2025-12-04T10:08:48.0337381Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0337615Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0337819Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0338027Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0338217Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0338397Z raise self.failureException(msg) 2025-12-04T10:08:48.0338528Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0338606Z 2025-12-04T10:08:48.0338686Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0338951Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0339139Z 2025-12-04T10:08:48.0339231Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0339453Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0339656Z Traceback (most recent call last): 2025-12-04T10:08:48.0339872Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0340123Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0340329Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0340533Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0340724Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0340906Z raise self.failureException(msg) 2025-12-04T10:08:48.0341036Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0341115Z 2025-12-04T10:08:48.0341196Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0341464Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0341652Z 2025-12-04T10:08:48.0341745Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0341967Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0342152Z Traceback (most recent call last): 2025-12-04T10:08:48.0342385Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0342615Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0342811Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0343006Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0343186Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0343357Z raise self.failureException(msg) 2025-12-04T10:08:48.0343494Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0343571Z 2025-12-04T10:08:48.0343646Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0343902Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0344091Z 2025-12-04T10:08:48.0344177Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0344390Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0344562Z Traceback (most recent call last): 2025-12-04T10:08:48.0344769Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0344989Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0345184Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0345379Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0345560Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0345732Z raise self.failureException(msg) 2025-12-04T10:08:48.0345852Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0345931Z 2025-12-04T10:08:48.0346007Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0346262Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0346457Z 2025-12-04T10:08:48.0346542Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0346833Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0347006Z Traceback (most recent call last): 2025-12-04T10:08:48.0347213Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0347435Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0347630Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0347852Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0348048Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0348222Z raise self.failureException(msg) 2025-12-04T10:08:48.0348343Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0348419Z 2025-12-04T10:08:48.0348494Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0348750Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0348934Z 2025-12-04T10:08:48.0349020Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0349234Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0349411Z Traceback (most recent call last): 2025-12-04T10:08:48.0349620Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0349840Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0350060Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0350257Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0350437Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0350608Z raise self.failureException(msg) 2025-12-04T10:08:48.0350729Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0350805Z 2025-12-04T10:08:48.0350878Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0351133Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0351318Z 2025-12-04T10:08:48.0351432Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0351648Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0351819Z Traceback (most recent call last): 2025-12-04T10:08:48.0352030Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0352252Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0352448Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0352645Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0352826Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0352998Z raise self.failureException(msg) 2025-12-04T10:08:48.0353119Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0353195Z 2025-12-04T10:08:48.0353271Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0353529Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0353713Z 2025-12-04T10:08:48.0353801Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0354016Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0354188Z Traceback (most recent call last): 2025-12-04T10:08:48.0354393Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0354614Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0354809Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0355004Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0355186Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0355358Z raise self.failureException(msg) 2025-12-04T10:08:48.0355496Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0355573Z 2025-12-04T10:08:48.0355645Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0355915Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0356100Z 2025-12-04T10:08:48.0356185Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0356398Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0356569Z Traceback (most recent call last): 2025-12-04T10:08:48.0356819Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0357044Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0357242Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0357439Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0357619Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0357792Z raise self.failureException(msg) 2025-12-04T10:08:48.0357932Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0358009Z 2025-12-04T10:08:48.0358081Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0358339Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0358524Z 2025-12-04T10:08:48.0358609Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0358821Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0358992Z Traceback (most recent call last): 2025-12-04T10:08:48.0359218Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0359444Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0359638Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0359836Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0360017Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0360188Z raise self.failureException(msg) 2025-12-04T10:08:48.0360309Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0360386Z 2025-12-04T10:08:48.0360458Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0360713Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0360899Z 2025-12-04T10:08:48.0360985Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0361197Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0361370Z Traceback (most recent call last): 2025-12-04T10:08:48.0361577Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0361803Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0361998Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0362194Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0362373Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0362544Z raise self.failureException(msg) 2025-12-04T10:08:48.0362664Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0362740Z 2025-12-04T10:08:48.0362814Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0363072Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0363275Z 2025-12-04T10:08:48.0363360Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0363593Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0363765Z Traceback (most recent call last): 2025-12-04T10:08:48.0363972Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0364193Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0364390Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0364585Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0364764Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0364937Z raise self.failureException(msg) 2025-12-04T10:08:48.0365058Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0365137Z 2025-12-04T10:08:48.0365211Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0365467Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0365667Z 2025-12-04T10:08:48.0365755Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0365966Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0366138Z Traceback (most recent call last): 2025-12-04T10:08:48.0366346Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0366568Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0366804Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0367020Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0367203Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0367373Z raise self.failureException(msg) 2025-12-04T10:08:48.0367498Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0367573Z 2025-12-04T10:08:48.0367650Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0367909Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0368094Z 2025-12-04T10:08:48.0368181Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0368394Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0368565Z Traceback (most recent call last): 2025-12-04T10:08:48.0368774Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0368995Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0369193Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0369388Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0369575Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0369747Z raise self.failureException(msg) 2025-12-04T10:08:48.0369867Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0369945Z 2025-12-04T10:08:48.0370017Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0370274Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0370461Z 2025-12-04T10:08:48.0370546Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0370760Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0370958Z Traceback (most recent call last): 2025-12-04T10:08:48.0371166Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0371403Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0371600Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0371796Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0371978Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0372149Z raise self.failureException(msg) 2025-12-04T10:08:48.0372269Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0372345Z 2025-12-04T10:08:48.0372418Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0372676Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0372862Z 2025-12-04T10:08:48.0372953Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0373166Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0373340Z Traceback (most recent call last): 2025-12-04T10:08:48.0373562Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0373784Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0373979Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0374175Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0374356Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0374528Z raise self.failureException(msg) 2025-12-04T10:08:48.0374648Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0374739Z 2025-12-04T10:08:48.0374812Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0375074Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0375263Z 2025-12-04T10:08:48.0375349Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0375561Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0375732Z Traceback (most recent call last): 2025-12-04T10:08:48.0375939Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0376159Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0376358Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0376559Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0376788Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0376963Z raise self.failureException(msg) 2025-12-04T10:08:48.0377089Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0377168Z 2025-12-04T10:08:48.0377243Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0377503Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0377690Z 2025-12-04T10:08:48.0377779Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0377997Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0378171Z Traceback (most recent call last): 2025-12-04T10:08:48.0378379Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0378604Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0378807Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0379024Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0379207Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0379401Z raise self.failureException(msg) 2025-12-04T10:08:48.0379526Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0379602Z 2025-12-04T10:08:48.0379677Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0379936Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0380123Z 2025-12-04T10:08:48.0380211Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0380427Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0380603Z Traceback (most recent call last): 2025-12-04T10:08:48.0380813Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0381040Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0381240Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0381454Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0381806Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0381982Z raise self.failureException(msg) 2025-12-04T10:08:48.0382105Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0382182Z 2025-12-04T10:08:48.0382256Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0382515Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0382702Z 2025-12-04T10:08:48.0382809Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0383028Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0383202Z Traceback (most recent call last): 2025-12-04T10:08:48.0383417Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0383643Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0383840Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0384039Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0384221Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0384393Z raise self.failureException(msg) 2025-12-04T10:08:48.0384516Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0384594Z 2025-12-04T10:08:48.0384669Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0384928Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0385119Z 2025-12-04T10:08:48.0385205Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0385423Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0385598Z Traceback (most recent call last): 2025-12-04T10:08:48.0385807Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0386032Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0386230Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0386428Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0386610Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0386825Z raise self.failureException(msg) 2025-12-04T10:08:48.0386969Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0387048Z 2025-12-04T10:08:48.0387121Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0387402Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0387590Z 2025-12-04T10:08:48.0387677Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0387894Z ____ TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 ____ 2025-12-04T10:08:48.0388068Z Traceback (most recent call last): 2025-12-04T10:08:48.0388278Z File "/var/lib/jenkins/pytorch/test/test_linalg.py", line 6098, in test_ops_append_to_existing_file_tunableop 2025-12-04T10:08:48.0388502Z self.assertGreater(final_count, initial_count) 2025-12-04T10:08:48.0388702Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 1244, in assertGreater 2025-12-04T10:08:48.0388901Z self.fail(self._formatMessage(msg, standardMsg)) 2025-12-04T10:08:48.0389088Z File "/opt/conda/envs/py_3.10/lib/python3.10/unittest/case.py", line 675, in fail 2025-12-04T10:08:48.0389264Z raise self.failureException(msg) 2025-12-04T10:08:48.0389388Z AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0389484Z 2025-12-04T10:08:48.0389559Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0389818Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0390006Z 2025-12-04T10:08:48.0390093Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0390386Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/test_linalg/test_linalg-115bbfbfd2646928.xml - 2025-12-04T10:08:48.0390673Z =========================== short test summary info ============================ 2025-12-04T10:08:48.0390955Z FAILED [0.0448s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0391171Z 2025-12-04T10:08:48.0391246Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0391509Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0391698Z 2025-12-04T10:08:48.0391784Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0392080Z FAILED [0.0497s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0392294Z 2025-12-04T10:08:48.0392367Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0392628Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0392816Z 2025-12-04T10:08:48.0392903Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0393196Z FAILED [0.0025s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0393408Z 2025-12-04T10:08:48.0393485Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0393743Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0393929Z 2025-12-04T10:08:48.0394014Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0394306Z FAILED [0.0021s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0394517Z 2025-12-04T10:08:48.0394592Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0394847Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0395053Z 2025-12-04T10:08:48.0395139Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0395448Z FAILED [0.0022s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0395658Z 2025-12-04T10:08:48.0395737Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0396004Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0396189Z 2025-12-04T10:08:48.0396281Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0396584Z FAILED [0.0020s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0396833Z 2025-12-04T10:08:48.0396912Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0397177Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0397364Z 2025-12-04T10:08:48.0397470Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0397771Z FAILED [0.0021s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0397981Z 2025-12-04T10:08:48.0398062Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0398326Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0398514Z 2025-12-04T10:08:48.0398626Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0398930Z FAILED [0.0018s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0399145Z 2025-12-04T10:08:48.0399225Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0399492Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0399680Z 2025-12-04T10:08:48.0399770Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0400065Z FAILED [0.0020s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0400276Z 2025-12-04T10:08:48.0400347Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0400608Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0400794Z 2025-12-04T10:08:48.0400879Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0401171Z FAILED [0.0018s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0401386Z 2025-12-04T10:08:48.0401460Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0401714Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0401899Z 2025-12-04T10:08:48.0401983Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0402278Z FAILED [0.0020s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0402493Z 2025-12-04T10:08:48.0402565Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0402828Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0403040Z 2025-12-04T10:08:48.0403125Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0403437Z FAILED [0.0018s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0403648Z 2025-12-04T10:08:48.0403721Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0403978Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0404164Z 2025-12-04T10:08:48.0404248Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0404542Z FAILED [0.0020s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0404757Z 2025-12-04T10:08:48.0404830Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0405089Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0405278Z 2025-12-04T10:08:48.0405378Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0405673Z FAILED [0.0018s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0405884Z 2025-12-04T10:08:48.0405955Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0406212Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0406401Z 2025-12-04T10:08:48.0406499Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0406833Z FAILED [0.0019s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0407044Z 2025-12-04T10:08:48.0407123Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0407386Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0407570Z 2025-12-04T10:08:48.0407656Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0407949Z FAILED [0.0017s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0408157Z 2025-12-04T10:08:48.0408234Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0408495Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0408682Z 2025-12-04T10:08:48.0408769Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0409065Z FAILED [0.0019s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0409276Z 2025-12-04T10:08:48.0409352Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0409609Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0409794Z 2025-12-04T10:08:48.0409881Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0410176Z FAILED [0.0017s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0410388Z 2025-12-04T10:08:48.0410463Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0410722Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0410927Z 2025-12-04T10:08:48.0411029Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0411324Z FAILED [0.0019s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0411536Z 2025-12-04T10:08:48.0411610Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0411870Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0412057Z 2025-12-04T10:08:48.0412141Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0412436Z FAILED [0.0018s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0412649Z 2025-12-04T10:08:48.0412720Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0412978Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0413180Z 2025-12-04T10:08:48.0413268Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0413561Z FAILED [0.0019s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0413773Z 2025-12-04T10:08:48.0413846Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0414108Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0414296Z 2025-12-04T10:08:48.0414396Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0414695Z FAILED [0.0017s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0414909Z 2025-12-04T10:08:48.0414984Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0415243Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0415431Z 2025-12-04T10:08:48.0415515Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0415808Z FAILED [0.0019s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0416017Z 2025-12-04T10:08:48.0416089Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0416349Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0416538Z 2025-12-04T10:08:48.0416623Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0416963Z FAILED [0.0017s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0417178Z 2025-12-04T10:08:48.0417251Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0417513Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0417702Z 2025-12-04T10:08:48.0417787Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0418080Z FAILED [0.0020s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0418294Z 2025-12-04T10:08:48.0418372Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0418650Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0418835Z 2025-12-04T10:08:48.0418943Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0419240Z FAILED [0.0018s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0419450Z 2025-12-04T10:08:48.0419530Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0419794Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0419981Z 2025-12-04T10:08:48.0420072Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0420373Z FAILED [0.0019s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0420586Z 2025-12-04T10:08:48.0420666Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0420930Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0421139Z 2025-12-04T10:08:48.0421232Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0421532Z FAILED [0.0017s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0421742Z 2025-12-04T10:08:48.0421822Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0422089Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0422280Z 2025-12-04T10:08:48.0422418Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0422722Z FAILED [0.0019s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0422932Z 2025-12-04T10:08:48.0423012Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0423277Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0423464Z 2025-12-04T10:08:48.0423554Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0423855Z FAILED [0.0017s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0424071Z 2025-12-04T10:08:48.0424145Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0424413Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0424607Z 2025-12-04T10:08:48.0424694Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0424997Z FAILED [0.0021s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0425214Z 2025-12-04T10:08:48.0425289Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0425556Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0425751Z 2025-12-04T10:08:48.0425840Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0426144Z FAILED [0.0010s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0426361Z 2025-12-04T10:08:48.0426436Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0426720Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0426950Z 2025-12-04T10:08:48.0427052Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0427357Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0427574Z 2025-12-04T10:08:48.0427648Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0427912Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0428102Z 2025-12-04T10:08:48.0428188Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0428487Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0428702Z 2025-12-04T10:08:48.0428776Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0429040Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0429253Z 2025-12-04T10:08:48.0429340Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0429639Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0429853Z 2025-12-04T10:08:48.0429931Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0430188Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0430378Z 2025-12-04T10:08:48.0430479Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0430774Z FAILED [0.0008s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0430984Z 2025-12-04T10:08:48.0431061Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0431326Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0431513Z 2025-12-04T10:08:48.0431604Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0431905Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0432116Z 2025-12-04T10:08:48.0432194Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0432460Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0432648Z 2025-12-04T10:08:48.0432739Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0433039Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0433252Z 2025-12-04T10:08:48.0433330Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0433597Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0433783Z 2025-12-04T10:08:48.0433874Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0434171Z FAILED [0.0008s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0434380Z 2025-12-04T10:08:48.0434457Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0434734Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0434920Z 2025-12-04T10:08:48.0435022Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0435317Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0435524Z 2025-12-04T10:08:48.0435598Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0435856Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0436044Z 2025-12-04T10:08:48.0436129Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0436425Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0436635Z 2025-12-04T10:08:48.0436707Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0436995Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0437203Z 2025-12-04T10:08:48.0437287Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0437582Z FAILED [0.0008s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0437792Z 2025-12-04T10:08:48.0437864Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0438124Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0438315Z 2025-12-04T10:08:48.0438415Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0438718Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0438928Z 2025-12-04T10:08:48.0439002Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0439261Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0439449Z 2025-12-04T10:08:48.0439533Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0439829Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0440039Z 2025-12-04T10:08:48.0440111Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0440370Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0440559Z 2025-12-04T10:08:48.0440643Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0440937Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0441149Z 2025-12-04T10:08:48.0441222Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0441482Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0441669Z 2025-12-04T10:08:48.0441757Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0442051Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0442263Z 2025-12-04T10:08:48.0442339Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0442617Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0442803Z 2025-12-04T10:08:48.0442906Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0443201Z FAILED [0.0007s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0443409Z 2025-12-04T10:08:48.0443485Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0443742Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0443926Z 2025-12-04T10:08:48.0444013Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0444306Z FAILED [0.0009s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0444515Z 2025-12-04T10:08:48.0444590Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0444849Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0445048Z 2025-12-04T10:08:48.0445138Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0445432Z FAILED [0.0008s] test_linalg.py::TestLinalgCUDA::test_ops_append_to_existing_file_tunableop_cuda_float32 - AssertionError: 1 not greater than 1 2025-12-04T10:08:48.0445639Z 2025-12-04T10:08:48.0445713Z To execute this test, run the following from the base repo dir: 2025-12-04T10:08:48.0445970Z PYTORCH_TEST_WITH_ROCM=1 python test/test_linalg.py TestLinalgCUDA.test_ops_append_to_existing_file_tunableop_cuda_float32 2025-12-04T10:08:48.0446156Z 2025-12-04T10:08:48.0446255Z This message can be suppressed by setting PYTORCH_PRINT_REPRO_ON_FAILURE=0 2025-12-04T10:08:48.0446442Z ======================== 49 failed, 1 passed in 39.16s ========================= 2025-12-04T10:08:48.0446541Z 2025-12-04T10:08:48.0446670Z FINISHED PRINTING LOG FILE of test_linalg 2/2 (test/test-reports/test_linalg_2.2_4f98d9a906e7bcd5_.log) 2025-12-04T10:08:48.0446867Z 2025-12-04T10:08:48.0446963Z Finished test_linalg 2/2 ... [2025-12-04 10:08:48.021283][2226752.478601191], took 0.72min 2025-12-04T10:08:48.0447323Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:08:48.0447690Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:08:48.0447910Z GITHUB_RUN_ID, GITHUB_RUN_ATTEMPT, or ARTIFACTS_FILE_SUFFIX not set, not uploading 2025-12-04T10:08:48.0448090Z Uploading artifacts took 0.00 seconds 2025-12-04T10:08:48.0448212Z test_linalg 2/2 failed! 2025-12-04T10:08:48.0448373Z Running nn/test_packed_sequence 1/1 ... [2025-12-04 10:08:48.034825][2226752.492147027] 2025-12-04T10:08:48.0448554Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:08:48.0449004Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_packed_sequence.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:08:48.035090] 2025-12-04T10:08:50.0953914Z 2025-12-04T10:08:50.0955064Z nn/test_packed_sequence 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_packed_sequence_1.1_3fe19267f91340cd_.log 2025-12-04T10:08:50.0955670Z Running 0 items in this shard: 2025-12-04T10:08:50.0955831Z 2025-12-04T10:08:50.0956059Z Finished nn/test_packed_sequence 1/1 ... [2025-12-04 10:08:50.094627][2226754.551945516], took 0.03min 2025-12-04T10:08:50.0965748Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:08:50.1083918Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:08:50.1085684Z Running test_numa_binding 1/1 ... [2025-12-04 10:08:50.108414][2226754.565736138] 2025-12-04T10:08:50.1086176Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:08:50.1089226Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_numa_binding.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:08:50.108703] 2025-12-04T10:08:52.2079060Z 2025-12-04T10:08:52.2080035Z test_numa_binding 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_numa_binding_1.1_67e6d51ed3d81004_.log 2025-12-04T10:08:52.2080658Z Running 0 items in this shard: 2025-12-04T10:08:52.2080864Z 2025-12-04T10:08:52.2081100Z Finished test_numa_binding 1/1 ... [2025-12-04 10:08:52.207702][2226756.665019221], took 0.03min 2025-12-04T10:08:52.2099818Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:08:52.2221972Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:08:52.2222357Z Running test_pruning_op 1/1 ... [2025-12-04 10:08:52.221612][2226756.678933821] 2025-12-04T10:08:52.2222646Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:08:52.2223393Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_pruning_op.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:08:52.221886] 2025-12-04T10:08:55.1129415Z 2025-12-04T10:08:55.1133273Z test_pruning_op 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_pruning_op_1.1_7f3c5ebb7ff44d0a_.log 2025-12-04T10:08:55.1134027Z Running 0 items in this shard: 2025-12-04T10:08:55.1134248Z 2025-12-04T10:08:55.1134546Z Finished test_pruning_op 1/1 ... [2025-12-04 10:08:55.112624][2226759.569943814], took 0.05min 2025-12-04T10:08:55.1141650Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:08:55.1257745Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:08:55.1262905Z Running test_jit_fuser_te 1/1 ... [2025-12-04 10:08:55.125966][2226759.583287985] 2025-12-04T10:08:55.1263113Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:08:55.1264534Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_jit_fuser_te.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:08:55.126248] 2025-12-04T10:09:06.9641785Z 2025-12-04T10:09:06.9642643Z test_jit_fuser_te 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jit_fuser_te_1.1_e59c69c597ca5a1f_.log 2025-12-04T10:09:06.9663446Z Running 150 items in this shard: test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserStatic::test_torch_to, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_inlined_optimized_graph, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check, test/test_jit_fuser_te.py::TestTEFuserDynamic::test_skip_grad_in_check 2025-12-04T10:09:06.9683300Z 2025-12-04T10:09:06.9683411Z Finished test_jit_fuser_te 1/1 ... [2025-12-04 10:09:06.963108][2226771.420427613], took 0.20min 2025-12-04T10:09:06.9688435Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:09:06.9757856Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:09:06.9760190Z Running nn/test_dropout 1/1 ... [2025-12-04 10:09:06.975916][2226771.433238425] 2025-12-04T10:09:06.9760372Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:09:06.9762866Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_dropout.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:09:06.976189] 2025-12-04T10:09:09.6377745Z 2025-12-04T10:09:09.6378705Z nn/test_dropout 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_dropout_1.1_32420f04c051955c_.log 2025-12-04T10:09:09.6379260Z Running 0 items in this shard: 2025-12-04T10:09:09.6379429Z 2025-12-04T10:09:09.6379678Z Finished nn/test_dropout 1/1 ... [2025-12-04 10:09:09.637569][2226774.09488769], took 0.04min 2025-12-04T10:09:09.6394385Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:09:09.6504960Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:09:09.6507025Z Running functorch/dim/test_split 1/1 ... [2025-12-04 10:09:09.650538][2226774.10786058] 2025-12-04T10:09:09.6507358Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:09:09.6509202Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/dim/test_split.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:09:09.650784] 2025-12-04T10:09:12.4564196Z 2025-12-04T10:09:12.4565165Z functorch/dim/test_split 1/1 was successful, full logs can be found in artifacts with path test/test-reports/functorch.dim.test_split_1.1_60c2d30e62468794_.log 2025-12-04T10:09:12.4566137Z Running 0 items in this shard: 2025-12-04T10:09:12.4566235Z 2025-12-04T10:09:12.4566386Z Finished functorch/dim/test_split 1/1 ... [2025-12-04 10:09:12.456232][2226776.913550688], took 0.05min 2025-12-04T10:09:12.4582088Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:09:12.4693687Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:09:12.4696499Z Running torch_np/numpy_tests/lib/test_type_check 1/1 ... [2025-12-04 10:09:12.469487][2226776.926809642] 2025-12-04T10:09:12.4697014Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:09:12.4699113Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/lib/test_type_check.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:09:12.469762] 2025-12-04T10:09:14.5699512Z 2025-12-04T10:09:14.5704028Z torch_np/numpy_tests/lib/test_type_check 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.lib.test_type_check_1.1_982556cecf7a6d46_.log 2025-12-04T10:09:14.5704401Z Running 0 items in this shard: 2025-12-04T10:09:14.5704481Z 2025-12-04T10:09:14.5704629Z Finished torch_np/numpy_tests/lib/test_type_check 1/1 ... [2025-12-04 10:09:14.569753][2226779.027073193], took 0.04min 2025-12-04T10:09:14.5713661Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:09:14.5823280Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:09:14.5825492Z Running cpp_extensions/test_libtorch_agnostic 1/1 ... [2025-12-04 10:09:14.582441][2226779.039765379] 2025-12-04T10:09:14.5825732Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:09:14.5827911Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'cpp_extensions/test_libtorch_agnostic.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:09:14.582663] 2025-12-04T10:09:34.2087881Z 2025-12-04T10:09:34.2089047Z cpp_extensions/test_libtorch_agnostic 1/1 was successful, full logs can be found in artifacts with path test/test-reports/cpp_extensions.test_libtorch_agnostic_1.1_60964b682b62f3f4_.log 2025-12-04T10:09:34.2103308Z Running 50 items in this shard: test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda, test/cpp_extensions/test_libtorch_agnostic.py::TestLibtorchAgnosticCUDA::test_mv_tensor_accessor_cuda 2025-12-04T10:09:34.2112013Z 2025-12-04T10:09:34.2112166Z Finished cpp_extensions/test_libtorch_agnostic 1/1 ... [2025-12-04 10:09:34.208725][2226798.666044312], took 0.33min 2025-12-04T10:09:34.2112617Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:09:34.2213383Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:09:34.2213807Z Running test_cpp_extensions_stream_and_event 1/1 ... [2025-12-04 10:09:34.221250][2226798.678574243] 2025-12-04T10:09:34.2214022Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:09:34.2216845Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_cpp_extensions_stream_and_event.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:09:34.221483] 2025-12-04T10:09:37.1465005Z 2025-12-04T10:09:37.1466256Z test_cpp_extensions_stream_and_event 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_cpp_extensions_stream_and_event_1.1_420b8a8c52a865d3_.log 2025-12-04T10:09:37.1467224Z Running 0 items in this shard: 2025-12-04T10:09:37.1467404Z 2025-12-04T10:09:37.1467722Z Finished test_cpp_extensions_stream_and_event 1/1 ... [2025-12-04 10:09:37.146204][2226801.603521534], took 0.05min 2025-12-04T10:09:37.1488111Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:09:37.1593689Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:09:37.1595652Z Running profiler/test_execution_trace 1/1 ... [2025-12-04 10:09:37.159412][2226801.616735731] 2025-12-04T10:09:37.1595999Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:09:37.1597976Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'profiler/test_execution_trace.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:09:37.159637] 2025-12-04T10:09:40.5182521Z 2025-12-04T10:09:40.5183346Z profiler/test_execution_trace 1/1 was successful, full logs can be found in artifacts with path test/test-reports/profiler.test_execution_trace_1.1_e6ebb9f74c7c6882_.log 2025-12-04T10:09:40.5183721Z Running 0 items in this shard: 2025-12-04T10:09:40.5183800Z 2025-12-04T10:09:40.5183929Z Finished profiler/test_execution_trace 1/1 ... [2025-12-04 10:09:40.517984][2226804.975302745], took 0.06min 2025-12-04T10:09:40.5196599Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:09:40.5311333Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:09:40.5313821Z Running test_dispatch 1/1 ... [2025-12-04 10:09:40.531244][2226804.988565522] 2025-12-04T10:09:40.5314012Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:09:40.5316553Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_dispatch.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:09:40.531537] 2025-12-04T10:09:42.6863475Z 2025-12-04T10:09:42.6864228Z test_dispatch 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_dispatch_1.1_6b6050e5ce8b002f_.log 2025-12-04T10:09:42.6864517Z Running 0 items in this shard: 2025-12-04T10:09:42.6864597Z 2025-12-04T10:09:42.6864716Z Finished test_dispatch 1/1 ... [2025-12-04 10:09:42.686036][2226807.143354604], took 0.04min 2025-12-04T10:09:42.6879522Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:09:42.6988504Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:09:42.6991040Z Running test_datapipe 1/1 ... [2025-12-04 10:09:42.698946][2226807.156269898] 2025-12-04T10:09:42.6994431Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:09:42.6995250Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_datapipe.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:09:42.699169] 2025-12-04T10:09:44.8320508Z 2025-12-04T10:09:44.8321291Z test_datapipe 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_datapipe_1.1_7e65cb00ae4165ff_.log 2025-12-04T10:09:44.8321585Z Running 0 items in this shard: 2025-12-04T10:09:44.8321666Z 2025-12-04T10:09:44.8321774Z Finished test_datapipe 1/1 ... [2025-12-04 10:09:44.831809][2226809.289130448], took 0.04min 2025-12-04T10:09:44.8336486Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:09:44.8439752Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:09:44.8442111Z Running test_ops 1/5 ... [2025-12-04 10:09:44.844076][2226809.301399634] 2025-12-04T10:09:44.8446610Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:09:44.8447515Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_ops.py', '--shard-id=1', '--num-shards=5', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:09:44.844294] 2025-12-04T10:10:24.9314795Z 2025-12-04T10:10:24.9338508Z test_ops 1/5 was successful, full logs can be found in artifacts with path test/test-reports/test_ops_1.5_39bb37100ef7109a_.log 2025-12-04T10:10:24.9345500Z Running 50 items in this shard: test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32, test/test_ops.py::TestCommonCUDA::test_compare_cpu_nn_functional_conv1d_cuda_float32 2025-12-04T10:10:24.9351466Z 2025-12-04T10:10:24.9351568Z Finished test_ops 1/5 ... [2025-12-04 10:10:24.931257][2226849.388575483], took 0.67min 2025-12-04T10:10:24.9417406Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:10:24.9483348Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:10:24.9505398Z Running torch_np/numpy_tests/linalg/test_linalg 1/1 ... [2025-12-04 10:10:24.948538][2226849.405853278] 2025-12-04T10:10:24.9505630Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:10:24.9506119Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/linalg/test_linalg.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:10:24.948830] 2025-12-04T10:10:28.2323227Z 2025-12-04T10:10:28.2323992Z torch_np/numpy_tests/linalg/test_linalg 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.linalg.test_linalg_1.1_1a4273e0c32e8af5_.log 2025-12-04T10:10:28.2324392Z Running 0 items in this shard: 2025-12-04T10:10:28.2324478Z 2025-12-04T10:10:28.2324619Z Finished torch_np/numpy_tests/linalg/test_linalg 1/1 ... [2025-12-04 10:10:28.232069][2226852.689390585], took 0.05min 2025-12-04T10:10:28.2336539Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:10:28.2453301Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:10:28.2454408Z Running nn/test_module_hooks 1/1 ... [2025-12-04 10:10:28.245200][2226852.702522029] 2025-12-04T10:10:28.2454658Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:10:28.2456002Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_module_hooks.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:10:28.245494] 2025-12-04T10:10:30.7207783Z 2025-12-04T10:10:30.7208961Z nn/test_module_hooks 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_module_hooks_1.1_3c21e595736fdd38_.log 2025-12-04T10:10:30.7209707Z Running 0 items in this shard: 2025-12-04T10:10:30.7209790Z 2025-12-04T10:10:30.7209903Z Finished nn/test_module_hooks 1/1 ... [2025-12-04 10:10:30.716773][2226855.174094319], took 0.04min 2025-12-04T10:10:30.7210340Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:10:30.7348555Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:10:30.7378635Z Running torch_np/numpy_tests/lib/test_twodim_base 1/1 ... [2025-12-04 10:10:30.734960][2226855.192277007] 2025-12-04T10:10:30.7379217Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:10:30.7382804Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/lib/test_twodim_base.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:10:30.735233] 2025-12-04T10:10:33.1398917Z 2025-12-04T10:10:33.1399741Z torch_np/numpy_tests/lib/test_twodim_base 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.lib.test_twodim_base_1.1_65510aa3c522166e_.log 2025-12-04T10:10:33.1400620Z Running 0 items in this shard: 2025-12-04T10:10:33.1403513Z 2025-12-04T10:10:33.1403681Z Finished torch_np/numpy_tests/lib/test_twodim_base 1/1 ... [2025-12-04 10:10:33.136005][2226857.593326506], took 0.04min 2025-12-04T10:10:33.1404095Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:10:33.1517711Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:10:33.1518851Z Running profiler/test_memory_profiler 1/1 ... [2025-12-04 10:10:33.148468][2226857.605790552] 2025-12-04T10:10:33.1519113Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:10:33.1519638Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'profiler/test_memory_profiler.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:10:33.148698] 2025-12-04T10:10:36.0579052Z 2025-12-04T10:10:36.0579965Z profiler/test_memory_profiler 1/1 was successful, full logs can be found in artifacts with path test/test-reports/profiler.test_memory_profiler_1.1_0502a1b3a3dd5af3_.log 2025-12-04T10:10:36.0588996Z Running 50 items in this shard: test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none, test/profiler/test_memory_profiler.py::TestIdentifyGradients::test_extract_gradients_from_optimizer_set_to_none 2025-12-04T10:10:36.0596988Z 2025-12-04T10:10:36.0607718Z Finished profiler/test_memory_profiler 1/1 ... [2025-12-04 10:10:36.053138][2226860.510458436], took 0.05min 2025-12-04T10:10:36.0608249Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:10:36.0717869Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:10:36.0719326Z Running test_serialization 1/1 ... [2025-12-04 10:10:36.067741][2226860.525063162] 2025-12-04T10:10:36.0719986Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:10:36.0720589Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_serialization.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:10:36.067962] 2025-12-04T10:10:39.9174452Z 2025-12-04T10:10:39.9175156Z test_serialization 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_serialization_1.1_fa5d8ef563105b99_.log 2025-12-04T10:10:39.9180441Z Running 0 items in this shard: 2025-12-04T10:10:39.9180544Z 2025-12-04T10:10:39.9180661Z Finished test_serialization 1/1 ... [2025-12-04 10:10:39.917225][2226864.374546059], took 0.06min 2025-12-04T10:10:39.9190164Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:10:39.9297204Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:10:39.9299754Z Running test_indexing 1/1 ... [2025-12-04 10:10:39.929810][2226864.387133003] 2025-12-04T10:10:39.9299956Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:10:39.9301938Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_indexing.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:10:39.930040] 2025-12-04T10:10:43.7060792Z 2025-12-04T10:10:43.7062215Z test_indexing 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_indexing_1.1_5dcc17012fa52eea_.log 2025-12-04T10:10:43.7062524Z Running 0 items in this shard: 2025-12-04T10:10:43.7062610Z 2025-12-04T10:10:43.7062719Z Finished test_indexing 1/1 ... [2025-12-04 10:10:43.705855][2226868.163175799], took 0.06min 2025-12-04T10:10:43.7074679Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:10:43.7183668Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:10:43.7185831Z Running torch_np/numpy_tests/fft/test_pocketfft 1/1 ... [2025-12-04 10:10:43.718473][2226868.175795123] 2025-12-04T10:10:43.7187121Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:10:43.7189059Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/fft/test_pocketfft.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:10:43.718757] 2025-12-04T10:10:45.9370641Z 2025-12-04T10:10:45.9371314Z torch_np/numpy_tests/fft/test_pocketfft 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.fft.test_pocketfft_1.1_0f5d094bb010a9b9_.log 2025-12-04T10:10:45.9371670Z Running 0 items in this shard: 2025-12-04T10:10:45.9371755Z 2025-12-04T10:10:45.9371967Z Finished torch_np/numpy_tests/fft/test_pocketfft 1/1 ... [2025-12-04 10:10:45.936987][2226870.394306932], took 0.04min 2025-12-04T10:10:45.9385583Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:10:45.9494985Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:10:45.9497674Z Running functorch/test_ops 2/4 ... [2025-12-04 10:10:45.949643][2226870.406963935] 2025-12-04T10:10:45.9498591Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:10:45.9503966Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'functorch/test_ops.py', '--shard-id=2', '--num-shards=4', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:10:45.949969] 2025-12-04T10:10:59.8914625Z 2025-12-04T10:10:59.8915475Z functorch/test_ops 2/4 was successful, full logs can be found in artifacts with path test/test-reports/functorch.test_ops_2.4_f0fe1e611603c302_.log 2025-12-04T10:10:59.8916070Z Running 0 items in this shard: 2025-12-04T10:10:59.8916231Z 2025-12-04T10:10:59.8916461Z Finished functorch/test_ops 2/4 ... [2025-12-04 10:10:59.891148][2226884.348469524], took 0.23min 2025-12-04T10:10:59.8932008Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:10:59.9037272Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:10:59.9041346Z Running torch_np/numpy_tests/core/test_dlpack 1/1 ... [2025-12-04 10:10:59.903785][2226884.361107379] 2025-12-04T10:10:59.9041737Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:10:59.9042558Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/core/test_dlpack.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:10:59.904026] 2025-12-04T10:11:02.2648041Z 2025-12-04T10:11:02.2649244Z torch_np/numpy_tests/core/test_dlpack 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.core.test_dlpack_1.1_44f9867072463394_.log 2025-12-04T10:11:02.2649626Z Running 0 items in this shard: 2025-12-04T10:11:02.2649725Z 2025-12-04T10:11:02.2649870Z Finished torch_np/numpy_tests/core/test_dlpack 1/1 ... [2025-12-04 10:11:02.264447][2226886.721769167], took 0.04min 2025-12-04T10:11:02.2661743Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:11:02.2768282Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:11:02.2769326Z Running test_multiprocessing_spawn 1/1 ... [2025-12-04 10:11:02.276712][2226886.734036418] 2025-12-04T10:11:02.2769799Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:11:02.2771472Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_multiprocessing_spawn.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:11:02.276907] 2025-12-04T10:11:09.4895350Z 2025-12-04T10:11:09.4896503Z test_multiprocessing_spawn 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_multiprocessing_spawn_1.1_343936ef1a5615ca_.log 2025-12-04T10:11:09.4903366Z Running 0 items in this shard: 2025-12-04T10:11:09.4903598Z 2025-12-04T10:11:09.4903939Z Finished test_multiprocessing_spawn 1/1 ... [2025-12-04 10:11:09.489239][2226893.946561194], took 0.12min 2025-12-04T10:11:09.4907292Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:11:09.5014559Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:11:09.5015720Z Running test_mkldnn 1/1 ... [2025-12-04 10:11:09.501416][2226893.958739238] 2025-12-04T10:11:09.5015908Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:11:09.5017814Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_mkldnn.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:11:09.501622] 2025-12-04T10:11:12.9365872Z 2025-12-04T10:11:12.9367973Z test_mkldnn 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_mkldnn_1.1_c4555e1728c88aed_.log 2025-12-04T10:11:12.9368462Z Running 0 items in this shard: 2025-12-04T10:11:12.9368616Z 2025-12-04T10:11:12.9368799Z Finished test_mkldnn 1/1 ... [2025-12-04 10:11:12.936293][2226897.393614119], took 0.06min 2025-12-04T10:11:12.9381977Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:11:12.9487346Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:11:12.9489644Z Running torch_np/numpy_tests/lib/test_index_tricks 1/1 ... [2025-12-04 10:11:12.948753][2226897.406075377] 2025-12-04T10:11:12.9490022Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:11:12.9490875Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'torch_np/numpy_tests/lib/test_index_tricks.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:11:12.948964] 2025-12-04T10:11:15.1054896Z 2025-12-04T10:11:15.1056025Z torch_np/numpy_tests/lib/test_index_tricks 1/1 was successful, full logs can be found in artifacts with path test/test-reports/torch_np.numpy_tests.lib.test_index_tricks_1.1_4f255a8b12c0699c_.log 2025-12-04T10:11:15.1056684Z Running 0 items in this shard: 2025-12-04T10:11:15.1056980Z 2025-12-04T10:11:15.1057282Z Finished torch_np/numpy_tests/lib/test_index_tricks 1/1 ... [2025-12-04 10:11:15.105222][2226899.562543235], took 0.04min 2025-12-04T10:11:15.1070118Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:11:15.1176607Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:11:15.1180337Z Running test_jit_autocast 1/1 ... [2025-12-04 10:11:15.117716][2226899.575038344] 2025-12-04T10:11:15.1180637Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:11:15.1181380Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_jit_autocast.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:11:15.117953] 2025-12-04T10:11:18.5105697Z 2025-12-04T10:11:18.5106540Z test_jit_autocast 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_jit_autocast_1.1_ac36965e58d027ab_.log 2025-12-04T10:11:18.5107643Z Running 0 items in this shard: 2025-12-04T10:11:18.5107722Z 2025-12-04T10:11:18.5107831Z Finished test_jit_autocast 1/1 ... [2025-12-04 10:11:18.510309][2226902.967629831], took 0.06min 2025-12-04T10:11:18.5119871Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:11:18.5229007Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:11:18.5230946Z Running nn/test_pooling 1/1 ... [2025-12-04 10:11:18.522990][2226902.980310596] 2025-12-04T10:11:18.5233322Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:11:18.5233759Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'nn/test_pooling.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:11:18.523263] 2025-12-04T10:11:22.3329513Z 2025-12-04T10:11:22.3330568Z nn/test_pooling 1/1 was successful, full logs can be found in artifacts with path test/test-reports/nn.test_pooling_1.1_3afffd97baa83695_.log 2025-12-04T10:11:22.3331417Z Running 0 items in this shard: 2025-12-04T10:11:22.3332252Z 2025-12-04T10:11:22.3334484Z Finished nn/test_pooling 1/1 ... [2025-12-04 10:11:22.332602][2226906.789923497], took 0.06min 2025-12-04T10:11:22.3343294Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:11:22.3452750Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:11:22.3453080Z Running lazy/test_reuse_ir 1/1 ... [2025-12-04 10:11:22.345147][2226906.802469465] 2025-12-04T10:11:22.3453353Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:11:22.3456302Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'lazy/test_reuse_ir.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:11:22.345359] 2025-12-04T10:11:24.3021439Z 2025-12-04T10:11:24.3022443Z lazy/test_reuse_ir 1/1 was successful, full logs can be found in artifacts with path test/test-reports/lazy.test_reuse_ir_1.1_10efbfa2f031bf8f_.log 2025-12-04T10:11:24.3023210Z Running 0 items in this shard: 2025-12-04T10:11:24.3023444Z 2025-12-04T10:11:24.3023738Z Finished lazy/test_reuse_ir 1/1 ... [2025-12-04 10:11:24.301801][2226908.759123112], took 0.03min 2025-12-04T10:11:24.3035465Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:11:24.3146077Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:11:24.3148456Z Running test_functional_autograd_benchmark 1/1 ... [2025-12-04 10:11:24.314669][2226908.771986954] 2025-12-04T10:11:24.3148885Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:11:24.3151096Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_functional_autograd_benchmark.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:11:24.314961] 2025-12-04T10:11:26.2778614Z 2025-12-04T10:11:26.2779223Z test_functional_autograd_benchmark 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_functional_autograd_benchmark_1.1_3de346a51954cdf9_.log 2025-12-04T10:11:26.2779590Z Running 0 items in this shard: 2025-12-04T10:11:26.2779675Z 2025-12-04T10:11:26.2779818Z Finished test_functional_autograd_benchmark 1/1 ... [2025-12-04 10:11:26.277645][2226910.734964694], took 0.03min 2025-12-04T10:11:26.2798968Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:11:26.2905417Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:11:26.2907437Z Running cpp_extensions/libtorch_agnostic_2_10_extension/test_version_compatibility 1/1 ... [2025-12-04 10:11:26.290642][2226910.747965534] 2025-12-04T10:11:26.2907709Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:11:26.2909842Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'cpp_extensions/libtorch_agnostic_2_10_extension/test_version_compatibility.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:11:26.290869] 2025-12-04T10:11:28.3296571Z 2025-12-04T10:11:28.3297532Z cpp_extensions/libtorch_agnostic_2_10_extension/test_version_compatibility 1/1 was successful, full logs can be found in artifacts with path test/test-reports/cpp_extensions.libtorch_agnostic_2_10_extension.test_version_compatibility_1.1_e893628d85e67c29_.log 2025-12-04T10:11:28.3298052Z Running 0 items in this shard: 2025-12-04T10:11:28.3298132Z 2025-12-04T10:11:28.3299077Z Finished cpp_extensions/libtorch_agnostic_2_10_extension/test_version_compatibility 1/1 ... [2025-12-04 10:11:28.329397][2226912.786718887], took 0.03min 2025-12-04T10:11:28.3311833Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:11:28.3418101Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:11:28.3419997Z Running test_autoload_disable 1/1 ... [2025-12-04 10:11:28.341914][2226912.799238446] 2025-12-04T10:11:28.4927002Z Processing /var/lib/jenkins/pytorch/test/cpp_extensions 2025-12-04T10:11:31.6624502Z Preparing metadata (pyproject.toml) ... [?25l- done 2025-12-04T10:11:31.6635401Z [?25hBuilding wheels for collected packages: torch_test_cpp_extension 2025-12-04T10:14:38.8606133Z Building wheel for torch_test_cpp_extension (pyproject.toml) ... [?25l- \ | / - \ | / - \ | / - \ | / - done 2025-12-04T10:14:38.8684850Z [?25h Created wheel for torch_test_cpp_extension: filename=torch_test_cpp_extension-0.0.0-cp310-cp310-linux_x86_64.whl size=13062818 sha256=e14dd8880944658f476194581903fdf6deb6687e91dec613f33edf92f7a9e61a 2025-12-04T10:14:38.8685621Z Stored in directory: /tmp/pip-ephem-wheel-cache-ydpbe3oz/wheels/a9/2e/d7/a9e103243c0b754e2324c4ee6ddd055c388a2eefc520cfc979 2025-12-04T10:14:38.8700505Z Successfully built torch_test_cpp_extension 2025-12-04T10:14:39.0231647Z Installing collected packages: torch_test_cpp_extension 2025-12-04T10:14:39.1506283Z Successfully installed torch_test_cpp_extension-0.0.0 2025-12-04T10:14:40.6340957Z 2025-12-04T10:14:40.6341472Z Running tests... 2025-12-04T10:14:40.6341695Z ---------------------------------------------------------------------- 2025-12-04T10:14:40.8003915Z . 2025-12-04T10:14:40.8004070Z ---------------------------------------------------------------------- 2025-12-04T10:14:40.8004248Z Ran 1 test in 0.166s 2025-12-04T10:14:40.8004324Z 2025-12-04T10:14:40.8004361Z OK 2025-12-04T10:14:40.8004416Z 2025-12-04T10:14:40.8004472Z Generating XML reports... 2025-12-04T10:14:41.4403418Z Finished test_autoload_disable 1/1 ... [2025-12-04 10:14:41.439887][2227105.897187855], took 3.22min 2025-12-04T10:14:41.4417077Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:14:41.4523293Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:14:41.4525369Z Running test_cpp_extensions_aot_ninja 1/1 ... [2025-12-04 10:14:41.452432][2227105.909754873] 2025-12-04T10:14:41.6202935Z Processing /var/lib/jenkins/pytorch/test/cpp_extensions 2025-12-04T10:14:44.0300427Z Preparing metadata (pyproject.toml) ... [?25l- done 2025-12-04T10:14:44.0319377Z [?25hBuilding wheels for collected packages: torch_test_cpp_extension 2025-12-04T10:16:29.9478304Z Building wheel for torch_test_cpp_extension (pyproject.toml) ... [?25l- \ | / - \ | / - \ | / - \ | / done 2025-12-04T10:16:29.9548338Z [?25h Created wheel for torch_test_cpp_extension: filename=torch_test_cpp_extension-0.0.0-cp310-cp310-linux_x86_64.whl size=13063082 sha256=e19949dcad2dbfaf202b69510174a529069d3395568c163e5173909cc23c62da 2025-12-04T10:16:29.9550146Z Stored in directory: /tmp/pip-ephem-wheel-cache-ywiyw686/wheels/a9/2e/d7/a9e103243c0b754e2324c4ee6ddd055c388a2eefc520cfc979 2025-12-04T10:16:29.9564388Z Successfully built torch_test_cpp_extension 2025-12-04T10:16:30.1087235Z Installing collected packages: torch_test_cpp_extension 2025-12-04T10:16:30.2115690Z Successfully installed torch_test_cpp_extension-0.0.0 2025-12-04T10:16:30.4168118Z Processing /var/lib/jenkins/pytorch/test/cpp_extensions/no_python_abi_suffix_test 2025-12-04T10:16:31.4375452Z Preparing metadata (pyproject.toml) ... [?25l- done 2025-12-04T10:16:31.4389345Z [?25hBuilding wheels for collected packages: no_python_abi_suffix_test 2025-12-04T10:16:32.5959901Z Building wheel for no_python_abi_suffix_test (pyproject.toml) ... [?25l- \ done 2025-12-04T10:16:32.5969377Z [?25h Created wheel for no_python_abi_suffix_test: filename=no_python_abi_suffix_test-0.0.0-cp310-cp310-linux_x86_64.whl size=2944 sha256=affd280ed3daf3058ac3b879404a4a2bb42540b8af87a92d050f4efb26e6ce97 2025-12-04T10:16:32.5969879Z Stored in directory: /tmp/pip-ephem-wheel-cache-tb45747d/wheels/01/96/31/d3c48c51cc163420d8b3b57e95a07fda055add3ed0ea48001b 2025-12-04T10:16:32.5984461Z Successfully built no_python_abi_suffix_test 2025-12-04T10:16:32.7506273Z Installing collected packages: no_python_abi_suffix_test 2025-12-04T10:16:32.7538153Z Successfully installed no_python_abi_suffix_test-0.0.0 2025-12-04T10:16:32.8112327Z * Getting build dependencies for wheel... 2025-12-04T10:16:33.8681706Z /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/python_agnostic/csrc/ultra_norm.cu -> /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/python_agnostic/csrc/ultra_norm.cu [skipped, no changes] 2025-12-04T10:16:33.8683183Z Successfully preprocessed all matching files. 2025-12-04T10:16:33.8683567Z Total number of unsupported CUDA function calls: 0 2025-12-04T10:16:33.8683786Z 2025-12-04T10:16:33.8683797Z 2025-12-04T10:16:33.8683918Z Total number of replaced kernel launches: 0 2025-12-04T10:16:33.8824958Z running egg_info 2025-12-04T10:16:33.8850711Z creating python_agnostic.egg-info 2025-12-04T10:16:33.8852025Z writing python_agnostic.egg-info/PKG-INFO 2025-12-04T10:16:33.8854271Z writing dependency_links to python_agnostic.egg-info/dependency_links.txt 2025-12-04T10:16:33.8855433Z writing top-level names to python_agnostic.egg-info/top_level.txt 2025-12-04T10:16:33.8855825Z writing manifest file 'python_agnostic.egg-info/SOURCES.txt' 2025-12-04T10:16:33.9099264Z reading manifest file 'python_agnostic.egg-info/SOURCES.txt' 2025-12-04T10:16:33.9104956Z writing manifest file 'python_agnostic.egg-info/SOURCES.txt' 2025-12-04T10:16:34.3920604Z * Building wheel... 2025-12-04T10:16:35.3690978Z /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/python_agnostic/csrc/ultra_norm.cu -> /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/python_agnostic/csrc/ultra_norm.cu [skipped, no changes] 2025-12-04T10:16:35.3692433Z Successfully preprocessed all matching files. 2025-12-04T10:16:35.3692820Z Total number of unsupported CUDA function calls: 0 2025-12-04T10:16:35.3693059Z 2025-12-04T10:16:35.3693066Z 2025-12-04T10:16:35.3693194Z Total number of replaced kernel launches: 0 2025-12-04T10:16:35.3795000Z running bdist_wheel 2025-12-04T10:16:35.4049451Z running build 2025-12-04T10:16:35.4049664Z running build_ext 2025-12-04T10:16:35.4068336Z building 'python_agnostic._C' extension 2025-12-04T10:16:35.4072746Z creating /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/build/temp.linux-x86_64-cpython-310/python_agnostic/csrc 2025-12-04T10:16:47.0383198Z [1/1] /opt/rocm/bin/hipcc -I/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include -I/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -I/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/include/THH -I/opt/rocm/include -I/opt/conda/envs/py_3.10/include/python3.10 -c -c /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/python_agnostic/csrc/ultra_norm.cu -o /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/build/temp.linux-x86_64-cpython-310/python_agnostic/csrc/ultra_norm.o -D__HIP_PLATFORM_AMD__=1 -DUSE_ROCM=1 -DHIPBLAS_V2 -fPIC -DCUDA_HAS_FP16=1 -D__HIP_NO_HALF_OPERATORS__=1 -D__HIP_NO_HALF_CONVERSIONS__=1 -DHIP_ENABLE_WARP_SYNC_BUILTINS=1 -DTORCH_API_INCLUDE_EXTENSION_H -DPy_LIMITED_API=0x030A0000 -DTORCH_EXTENSION_NAME=_C --offload-arch=gfx90a --offload-arch=gfx942 --offload-arch=gfx950 --offload-arch=gfx1100 -fno-gpu-rdc -std=c++17 2025-12-04T10:16:47.0416014Z creating build/lib.linux-x86_64-cpython-310/python_agnostic 2025-12-04T10:16:47.0419683Z g++ -pthread -B /opt/conda/envs/py_3.10/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O2 -Wall -fPIC -O2 -isystem /opt/conda/envs/py_3.10/include -fPIC -O2 -isystem /opt/conda/envs/py_3.10/include -pthread -B /opt/conda/envs/py_3.10/compiler_compat -shared /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/build/temp.linux-x86_64-cpython-310/python_agnostic/csrc/ultra_norm.o -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -L/opt/rocm/lib -L/opt/rocm/hip/lib -lc10 -ltorch -ltorch_cpu -lamdhip64 -lc10_hip -ltorch_hip -o build/lib.linux-x86_64-cpython-310/python_agnostic/_C.so 2025-12-04T10:16:47.2159124Z installing to build/bdist.linux-x86_64/wheel 2025-12-04T10:16:47.2159352Z running install 2025-12-04T10:16:47.2180418Z running install_lib 2025-12-04T10:16:47.2201859Z creating build/bdist.linux-x86_64/wheel 2025-12-04T10:16:47.2202876Z creating build/bdist.linux-x86_64/wheel/python_agnostic 2025-12-04T10:16:47.2205300Z copying build/lib.linux-x86_64-cpython-310/python_agnostic/_C.so -> build/bdist.linux-x86_64/wheel/./python_agnostic 2025-12-04T10:16:47.2205641Z running install_egg_info 2025-12-04T10:16:47.2227051Z running egg_info 2025-12-04T10:16:47.2246923Z writing python_agnostic.egg-info/PKG-INFO 2025-12-04T10:16:47.2249950Z writing dependency_links to python_agnostic.egg-info/dependency_links.txt 2025-12-04T10:16:47.2252476Z writing top-level names to python_agnostic.egg-info/top_level.txt 2025-12-04T10:16:47.2275375Z reading manifest file 'python_agnostic.egg-info/SOURCES.txt' 2025-12-04T10:16:47.2280175Z writing manifest file 'python_agnostic.egg-info/SOURCES.txt' 2025-12-04T10:16:47.2281598Z Copying python_agnostic.egg-info to build/bdist.linux-x86_64/wheel/./python_agnostic-0.0-py3.10.egg-info 2025-12-04T10:16:47.2283701Z running install_scripts 2025-12-04T10:16:47.2330208Z creating build/bdist.linux-x86_64/wheel/python_agnostic-0.0.dist-info/WHEEL 2025-12-04T10:16:47.2332837Z creating '/var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/dist/.tmp-wb1kx_gc/python_agnostic-0.0-cp39-abi3-linux_x86_64.whl' and adding 'build/bdist.linux-x86_64/wheel' to it 2025-12-04T10:16:47.2346724Z adding 'python_agnostic/_C.so' 2025-12-04T10:16:47.2348469Z adding 'python_agnostic-0.0.dist-info/METADATA' 2025-12-04T10:16:47.2349017Z adding 'python_agnostic-0.0.dist-info/WHEEL' 2025-12-04T10:16:47.2349420Z adding 'python_agnostic-0.0.dist-info/top_level.txt' 2025-12-04T10:16:47.2349840Z adding 'python_agnostic-0.0.dist-info/RECORD' 2025-12-04T10:16:47.2350128Z removing build/bdist.linux-x86_64/wheel 2025-12-04T10:16:47.7079813Z Successfully built python_agnostic-0.0-cp39-abi3-linux_x86_64.whl 2025-12-04T10:16:47.8697052Z Processing /var/lib/jenkins/pytorch/test/cpp_extensions/libtorch_agnostic_2_9_extension 2025-12-04T10:16:49.2623257Z Preparing metadata (pyproject.toml) ... [?25l- done 2025-12-04T10:16:49.2646192Z [?25hRequirement already satisfied: torch in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from libtorch_agnostic_2_9==0.0) (2.10.0a0+gitffd9b0f) 2025-12-04T10:16:49.2660834Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (3.18.0) 2025-12-04T10:16:49.2662781Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (4.12.2) 2025-12-04T10:16:49.2664218Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (1.13.3) 2025-12-04T10:16:49.2666036Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (2.8.8) 2025-12-04T10:16:49.2667594Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (3.1.6) 2025-12-04T10:16:49.2669649Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (2025.10.0) 2025-12-04T10:16:49.2831637Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch->libtorch_agnostic_2_9==0.0) (1.3.0) 2025-12-04T10:16:49.2857027Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch->libtorch_agnostic_2_9==0.0) (3.0.3) 2025-12-04T10:16:49.2862196Z Building wheels for collected packages: libtorch_agnostic_2_9 2025-12-04T10:16:51.1002325Z Building wheel for libtorch_agnostic_2_9 (pyproject.toml) ... [?25l- \ done 2025-12-04T10:16:51.1009143Z [?25h Created wheel for libtorch_agnostic_2_9: filename=libtorch_agnostic_2_9-0.0-cp39-abi3-linux_x86_64.whl size=57687 sha256=bfec9d2b3b526151ce07144c572eba2c15500109c1d043318467eb6ed86ba056 2025-12-04T10:16:51.1010481Z Stored in directory: /tmp/pip-ephem-wheel-cache-_4_apzvw/wheels/5f/cf/13/81cd7305efc425780825ac0d330cbfca8f1ec0a935d4028550 2025-12-04T10:16:51.1025727Z Successfully built libtorch_agnostic_2_9 2025-12-04T10:16:51.2363230Z Installing collected packages: libtorch_agnostic_2_9 2025-12-04T10:16:51.2399936Z Successfully installed libtorch_agnostic_2_9-0.0 2025-12-04T10:16:51.4177354Z Processing /var/lib/jenkins/pytorch/test/cpp_extensions/libtorch_agnostic_2_10_extension 2025-12-04T10:16:52.8755194Z Preparing metadata (pyproject.toml) ... [?25l- done 2025-12-04T10:16:52.8775684Z [?25hRequirement already satisfied: torch in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from libtorch_agnostic_2_10==0.0) (2.10.0a0+gitffd9b0f) 2025-12-04T10:16:52.8789500Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (3.18.0) 2025-12-04T10:16:52.8790838Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (4.12.2) 2025-12-04T10:16:52.8792376Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (1.13.3) 2025-12-04T10:16:52.8794108Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (2.8.8) 2025-12-04T10:16:52.8796319Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (3.1.6) 2025-12-04T10:16:52.8797916Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (2025.10.0) 2025-12-04T10:16:52.8958787Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch->libtorch_agnostic_2_10==0.0) (1.3.0) 2025-12-04T10:16:52.8982296Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch->libtorch_agnostic_2_10==0.0) (3.0.3) 2025-12-04T10:16:52.8986926Z Building wheels for collected packages: libtorch_agnostic_2_10 2025-12-04T10:16:54.7413414Z Building wheel for libtorch_agnostic_2_10 (pyproject.toml) ... [?25l- \ done 2025-12-04T10:16:54.7417826Z [?25h Created wheel for libtorch_agnostic_2_10: filename=libtorch_agnostic_2_10-0.0-cp39-abi3-linux_x86_64.whl size=85717 sha256=94242e1bd9ed4d72e95eb259fdea4eeea92a2ff49dd263f3acc893c503ef1b4f 2025-12-04T10:16:54.7418925Z Stored in directory: /tmp/pip-ephem-wheel-cache-4beeii07/wheels/65/df/cc/92bed30159c7744af4875e08c646cf811fe190b78052f3234e 2025-12-04T10:16:54.7431254Z Successfully built libtorch_agnostic_2_10 2025-12-04T10:16:54.8740730Z Installing collected packages: libtorch_agnostic_2_10 2025-12-04T10:16:54.8776679Z Successfully installed libtorch_agnostic_2_10-0.0 2025-12-04T10:16:54.9053587Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:16:54.9056849Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_cpp_extensions_aot_ninja.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:16:54.905445] 2025-12-04T10:16:57.0201150Z 2025-12-04T10:16:57.0202323Z test_cpp_extensions_aot_ninja 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_cpp_extensions_aot_ninja_1.1_e288580916568e25_.log 2025-12-04T10:16:57.0203039Z Running 0 items in this shard: 2025-12-04T10:16:57.0203195Z 2025-12-04T10:16:57.0203814Z Finished test_cpp_extensions_aot_ninja 1/1 ... [2025-12-04 10:16:57.020003][2227241.477308855], took 2.26min 2025-12-04T10:16:57.0220471Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:16:57.0327714Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:16:57.0331397Z Running test_cpp_extensions_aot_no_ninja 1/1 ... [2025-12-04 10:16:57.032850][2227241.490174564] 2025-12-04T10:16:57.1891443Z Processing /var/lib/jenkins/pytorch/test/cpp_extensions 2025-12-04T10:16:59.7383000Z Preparing metadata (pyproject.toml) ... [?25l- done 2025-12-04T10:16:59.7399111Z [?25hBuilding wheels for collected packages: torch_test_cpp_extension 2025-12-04T10:17:04.6176451Z Building wheel for torch_test_cpp_extension (pyproject.toml) ... [?25l- \ | / - \ | / - \ | done 2025-12-04T10:17:04.6246712Z [?25h Created wheel for torch_test_cpp_extension: filename=torch_test_cpp_extension-0.0.0-cp310-cp310-linux_x86_64.whl size=13062818 sha256=f035337cd6e063e4e9b4f83ce058c54d7f86d61974d651d3afe01dc36a9a15e3 2025-12-04T10:17:04.6247310Z Stored in directory: /tmp/pip-ephem-wheel-cache-ngrhnv5y/wheels/a9/2e/d7/a9e103243c0b754e2324c4ee6ddd055c388a2eefc520cfc979 2025-12-04T10:17:04.6259289Z Successfully built torch_test_cpp_extension 2025-12-04T10:17:04.7764292Z Installing collected packages: torch_test_cpp_extension 2025-12-04T10:17:04.8803026Z Successfully installed torch_test_cpp_extension-0.0.0 2025-12-04T10:17:05.0781417Z Processing /var/lib/jenkins/pytorch/test/cpp_extensions/no_python_abi_suffix_test 2025-12-04T10:17:06.0622432Z Preparing metadata (pyproject.toml) ... [?25l- done 2025-12-04T10:17:06.0640225Z [?25hBuilding wheels for collected packages: no_python_abi_suffix_test 2025-12-04T10:17:07.1959111Z Building wheel for no_python_abi_suffix_test (pyproject.toml) ... [?25l- \ done 2025-12-04T10:17:07.1965648Z [?25h Created wheel for no_python_abi_suffix_test: filename=no_python_abi_suffix_test-0.0.0-cp310-cp310-linux_x86_64.whl size=2944 sha256=0e9bab9521786ed6381e50e4af0c485a81de41bd47fb20157a4d52c645fd5d3f 2025-12-04T10:17:07.1968510Z Stored in directory: /tmp/pip-ephem-wheel-cache-3nkfr2rm/wheels/01/96/31/d3c48c51cc163420d8b3b57e95a07fda055add3ed0ea48001b 2025-12-04T10:17:07.1981909Z Successfully built no_python_abi_suffix_test 2025-12-04T10:17:07.3482743Z Installing collected packages: no_python_abi_suffix_test 2025-12-04T10:17:07.3507426Z Successfully installed no_python_abi_suffix_test-0.0.0 2025-12-04T10:17:07.4063936Z * Getting build dependencies for wheel... 2025-12-04T10:17:08.4042897Z /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/python_agnostic/csrc/ultra_norm.cu -> /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/python_agnostic/csrc/ultra_norm.cu [skipped, no changes] 2025-12-04T10:17:08.4044562Z Successfully preprocessed all matching files. 2025-12-04T10:17:08.4045062Z Total number of unsupported CUDA function calls: 0 2025-12-04T10:17:08.4045347Z 2025-12-04T10:17:08.4045357Z 2025-12-04T10:17:08.4045504Z Total number of replaced kernel launches: 0 2025-12-04T10:17:08.4183569Z running egg_info 2025-12-04T10:17:08.4210904Z writing python_agnostic.egg-info/PKG-INFO 2025-12-04T10:17:08.4213487Z writing dependency_links to python_agnostic.egg-info/dependency_links.txt 2025-12-04T10:17:08.4214810Z writing top-level names to python_agnostic.egg-info/top_level.txt 2025-12-04T10:17:08.4444191Z reading manifest file 'python_agnostic.egg-info/SOURCES.txt' 2025-12-04T10:17:08.4450995Z writing manifest file 'python_agnostic.egg-info/SOURCES.txt' 2025-12-04T10:17:08.8338125Z * Building wheel... 2025-12-04T10:17:09.8497247Z /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/python_agnostic/csrc/ultra_norm.cu -> /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/python_agnostic/csrc/ultra_norm.cu [skipped, no changes] 2025-12-04T10:17:09.8499371Z Successfully preprocessed all matching files. 2025-12-04T10:17:09.8499818Z Total number of unsupported CUDA function calls: 0 2025-12-04T10:17:09.8500067Z 2025-12-04T10:17:09.8500072Z 2025-12-04T10:17:09.8500214Z Total number of replaced kernel launches: 0 2025-12-04T10:17:09.8600726Z running bdist_wheel 2025-12-04T10:17:09.8853348Z running build 2025-12-04T10:17:09.8853532Z running build_ext 2025-12-04T10:17:09.8867883Z building 'python_agnostic._C' extension 2025-12-04T10:17:09.9191541Z ninja: no work to do. 2025-12-04T10:17:09.9215200Z g++ -pthread -B /opt/conda/envs/py_3.10/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O2 -Wall -fPIC -O2 -isystem /opt/conda/envs/py_3.10/include -fPIC -O2 -isystem /opt/conda/envs/py_3.10/include -pthread -B /opt/conda/envs/py_3.10/compiler_compat -shared /var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/build/temp.linux-x86_64-cpython-310/python_agnostic/csrc/ultra_norm.o -L/opt/conda/envs/py_3.10/lib/python3.10/site-packages/torch/lib -L/opt/rocm/lib -L/opt/rocm/hip/lib -lc10 -ltorch -ltorch_cpu -lamdhip64 -lc10_hip -ltorch_hip -o build/lib.linux-x86_64-cpython-310/python_agnostic/_C.so 2025-12-04T10:17:10.0981152Z installing to build/bdist.linux-x86_64/wheel 2025-12-04T10:17:10.0981306Z running install 2025-12-04T10:17:10.1003879Z running install_lib 2025-12-04T10:17:10.1025326Z creating build/bdist.linux-x86_64/wheel 2025-12-04T10:17:10.1025724Z creating build/bdist.linux-x86_64/wheel/python_agnostic 2025-12-04T10:17:10.1026694Z copying build/lib.linux-x86_64-cpython-310/python_agnostic/_C.so -> build/bdist.linux-x86_64/wheel/./python_agnostic 2025-12-04T10:17:10.1027224Z running install_egg_info 2025-12-04T10:17:10.1050182Z running egg_info 2025-12-04T10:17:10.1069134Z writing python_agnostic.egg-info/PKG-INFO 2025-12-04T10:17:10.1071388Z writing dependency_links to python_agnostic.egg-info/dependency_links.txt 2025-12-04T10:17:10.1073456Z writing top-level names to python_agnostic.egg-info/top_level.txt 2025-12-04T10:17:10.1096020Z reading manifest file 'python_agnostic.egg-info/SOURCES.txt' 2025-12-04T10:17:10.1100725Z writing manifest file 'python_agnostic.egg-info/SOURCES.txt' 2025-12-04T10:17:10.1101499Z Copying python_agnostic.egg-info to build/bdist.linux-x86_64/wheel/./python_agnostic-0.0-py3.10.egg-info 2025-12-04T10:17:10.1103849Z running install_scripts 2025-12-04T10:17:10.1150815Z creating build/bdist.linux-x86_64/wheel/python_agnostic-0.0.dist-info/WHEEL 2025-12-04T10:17:10.1152698Z creating '/var/lib/jenkins/pytorch/test/cpp_extensions/python_agnostic_extension/dist/.tmp-_v659q_b/python_agnostic-0.0-cp39-abi3-linux_x86_64.whl' and adding 'build/bdist.linux-x86_64/wheel' to it 2025-12-04T10:17:10.1168974Z adding 'python_agnostic/_C.so' 2025-12-04T10:17:10.1170619Z adding 'python_agnostic-0.0.dist-info/METADATA' 2025-12-04T10:17:10.1170972Z adding 'python_agnostic-0.0.dist-info/WHEEL' 2025-12-04T10:17:10.1171304Z adding 'python_agnostic-0.0.dist-info/top_level.txt' 2025-12-04T10:17:10.1172051Z adding 'python_agnostic-0.0.dist-info/RECORD' 2025-12-04T10:17:10.1172312Z removing build/bdist.linux-x86_64/wheel 2025-12-04T10:17:10.5644243Z Successfully built python_agnostic-0.0-cp39-abi3-linux_x86_64.whl 2025-12-04T10:17:10.7216268Z Processing /var/lib/jenkins/pytorch/test/cpp_extensions/libtorch_agnostic_2_9_extension 2025-12-04T10:17:12.1193328Z Preparing metadata (pyproject.toml) ... [?25l- done 2025-12-04T10:17:12.1212701Z [?25hRequirement already satisfied: torch in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from libtorch_agnostic_2_9==0.0) (2.10.0a0+gitffd9b0f) 2025-12-04T10:17:12.1226127Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (3.18.0) 2025-12-04T10:17:12.1228900Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (4.12.2) 2025-12-04T10:17:12.1230954Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (1.13.3) 2025-12-04T10:17:12.1232319Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (2.8.8) 2025-12-04T10:17:12.1233729Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (3.1.6) 2025-12-04T10:17:12.1235506Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_9==0.0) (2025.10.0) 2025-12-04T10:17:12.1395653Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch->libtorch_agnostic_2_9==0.0) (1.3.0) 2025-12-04T10:17:12.1419442Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch->libtorch_agnostic_2_9==0.0) (3.0.3) 2025-12-04T10:17:12.1424473Z Building wheels for collected packages: libtorch_agnostic_2_9 2025-12-04T10:17:13.7599079Z Building wheel for libtorch_agnostic_2_9 (pyproject.toml) ... [?25l- \ done 2025-12-04T10:17:13.7607857Z [?25h Created wheel for libtorch_agnostic_2_9: filename=libtorch_agnostic_2_9-0.0-cp39-abi3-linux_x86_64.whl size=57687 sha256=ae5f9904f51a53f35ec1dbcb854ebc72e43c06c5d36aea10e125605812c84b2a 2025-12-04T10:17:13.7609478Z Stored in directory: /tmp/pip-ephem-wheel-cache-iv6d13pa/wheels/5f/cf/13/81cd7305efc425780825ac0d330cbfca8f1ec0a935d4028550 2025-12-04T10:17:13.7624851Z Successfully built libtorch_agnostic_2_9 2025-12-04T10:17:13.8991003Z Installing collected packages: libtorch_agnostic_2_9 2025-12-04T10:17:13.9030386Z Successfully installed libtorch_agnostic_2_9-0.0 2025-12-04T10:17:14.0875271Z Processing /var/lib/jenkins/pytorch/test/cpp_extensions/libtorch_agnostic_2_10_extension 2025-12-04T10:17:15.6036734Z Preparing metadata (pyproject.toml) ... [?25l- done 2025-12-04T10:17:15.6054519Z [?25hRequirement already satisfied: torch in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from libtorch_agnostic_2_10==0.0) (2.10.0a0+gitffd9b0f) 2025-12-04T10:17:15.6067764Z Requirement already satisfied: filelock in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (3.18.0) 2025-12-04T10:17:15.6069773Z Requirement already satisfied: typing-extensions>=4.10.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (4.12.2) 2025-12-04T10:17:15.6071405Z Requirement already satisfied: sympy>=1.13.3 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (1.13.3) 2025-12-04T10:17:15.6073200Z Requirement already satisfied: networkx>=2.5.1 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (2.8.8) 2025-12-04T10:17:15.6074497Z Requirement already satisfied: jinja2 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (3.1.6) 2025-12-04T10:17:15.6076212Z Requirement already satisfied: fsspec>=0.8.5 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from torch->libtorch_agnostic_2_10==0.0) (2025.10.0) 2025-12-04T10:17:15.6235976Z Requirement already satisfied: mpmath<1.4,>=1.1.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from sympy>=1.13.3->torch->libtorch_agnostic_2_10==0.0) (1.3.0) 2025-12-04T10:17:15.6259099Z Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/envs/py_3.10/lib/python3.10/site-packages (from jinja2->torch->libtorch_agnostic_2_10==0.0) (3.0.3) 2025-12-04T10:17:15.6263753Z Building wheels for collected packages: libtorch_agnostic_2_10 2025-12-04T10:17:17.3434640Z Building wheel for libtorch_agnostic_2_10 (pyproject.toml) ... [?25l- \ done 2025-12-04T10:17:17.3443064Z [?25h Created wheel for libtorch_agnostic_2_10: filename=libtorch_agnostic_2_10-0.0-cp39-abi3-linux_x86_64.whl size=85717 sha256=4fdee7c9149ef29a272b5781fa0b4858c5191f31199fcb93c27447ac1e3671c4 2025-12-04T10:17:17.3444508Z Stored in directory: /tmp/pip-ephem-wheel-cache-itbztdr3/wheels/65/df/cc/92bed30159c7744af4875e08c646cf811fe190b78052f3234e 2025-12-04T10:17:17.3459637Z Successfully built libtorch_agnostic_2_10 2025-12-04T10:17:17.4797093Z Installing collected packages: libtorch_agnostic_2_10 2025-12-04T10:17:17.4836925Z Successfully installed libtorch_agnostic_2_10-0.0 2025-12-04T10:17:17.5171321Z SCRIBE_GRAPHQL_ACCESS_TOKEN is NOT set 2025-12-04T10:17:17.5175834Z Executing ['/opt/conda/envs/py_3.10/bin/python', '-bb', 'test_cpp_extensions_aot_no_ninja.py', '--shard-id=1', '--num-shards=1', '-v', '-vv', '-rfEX', '-p', 'no:xdist', '--use-pytest', '--flake-finder', '--flake-runs=50', '--import-slow-tests', '--import-disabled-tests', '--rerun-disabled-tests'] ... [2025-12-04 10:17:17.517291] 2025-12-04T10:17:19.6587551Z 2025-12-04T10:17:19.6588415Z test_cpp_extensions_aot_no_ninja 1/1 was successful, full logs can be found in artifacts with path test/test-reports/test_cpp_extensions_aot_no_ninja_1.1_62a0c15346164cc5_.log 2025-12-04T10:17:19.6588756Z Running 0 items in this shard: 2025-12-04T10:17:19.6588837Z 2025-12-04T10:17:19.6588976Z Finished test_cpp_extensions_aot_no_ninja 1/1 ... [2025-12-04 10:17:19.658621][2227264.115942094], took 0.38min 2025-12-04T10:17:19.6603735Z Parsing testcases for test report: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/inductor.test_aot_inductor/inductor.test_aot_inductor-8ac8f4826ecf967f.xml 2025-12-04T10:17:19.6708324Z Failed to parse and upload json test reports: Unable to locate credentials 2025-12-04T10:17:22.0560717Z Running test batch 'tests to run' cost 2647.57 seconds 2025-12-04T10:17:22.0571807Z Emitting td_test_failure_stats_v2 2025-12-04T10:17:22.0576464Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764843442_6bfc25d2d0fa11f0b900da9c548f8d26 2025-12-04T10:17:23.0999734Z /var/lib/jenkins/pytorch/tools/stats/upload_metrics.py:156: UserWarning: Error uploading metric td_test_failure_stats_v2 to DynamoDB: Unable to locate credentials 2025-12-04T10:17:23.1000407Z warn(f"Error uploading metric {metric_name} to DynamoDB: {e}") 2025-12-04T10:17:23.1005034Z Emitting td_test_failure_stats_v2 2025-12-04T10:17:23.1005965Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764843443_6c9b521ad0fa11f0b900da9c548f8d26 2025-12-04T10:17:23.1038992Z Emitting td_test_failure_stats_v2 2025-12-04T10:17:23.1045607Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764843443_6c9bd226d0fa11f0b900da9c548f8d26 2025-12-04T10:17:23.1072400Z Emitting td_test_failure_stats_v2 2025-12-04T10:17:23.1072676Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764843443_6c9c5642d0fa11f0b900da9c548f8d26 2025-12-04T10:17:23.1106001Z Emitting td_test_failure_stats_v2 2025-12-04T10:17:23.1106292Z Writing 1 documents to S3 ossci-raw-job-status/ossci_uploaded_metrics/td_test_failure_stats_v2_1764843443_6c9cd98cd0fa11f0b900da9c548f8d26 2025-12-04T10:17:23.1135376Z inductor/test_aot_inductor 3/3 failed! 2025-12-04T10:17:23.1135537Z inductor/test_inductor_freezing 1/1 failed! 2025-12-04T10:17:23.1135684Z inductor/test_compile_subprocess 3/3 failed! 2025-12-04T10:17:23.1135927Z test_ci_sanity_check_fail 1/1 failed! 2025-12-04T10:17:23.1136048Z test_linalg 2/2 failed! 2025-12-04T10:17:23.8835418Z 2025-12-04T10:17:23.8835912Z real 44m13.013s 2025-12-04T10:17:23.8836082Z user 58m29.046s 2025-12-04T10:17:23.8836185Z sys 35m53.254s 2025-12-04T10:17:23.8836282Z + assert_git_not_dirty 2025-12-04T10:17:23.8836425Z + [[ linux-jammy-rocm-py3.10 != *rocm* ]] 2025-12-04T10:17:23.8836571Z + sccache_epilogue 2025-12-04T10:17:23.8836713Z + echo '::group::Sccache Compilation Log' 2025-12-04T10:17:23.8837214Z ##[group]Sccache Compilation Log 2025-12-04T10:17:23.8837692Z + echo '=================== sccache compilation log ===================' 2025-12-04T10:17:23.8837871Z =================== sccache compilation log =================== 2025-12-04T10:17:23.8838113Z + python /var/lib/jenkins/pytorch/.ci/pytorch/print_sccache_log.py /var/lib/jenkins/sccache_error.log 2025-12-04T10:17:23.8912664Z + echo '=========== If your build fails, please take a look at the log above for possible reasons ===========' 2025-12-04T10:17:23.8912936Z =========== If your build fails, please take a look at the log above for possible reasons =========== 2025-12-04T10:17:23.8913129Z + sccache --show-stats 2025-12-04T10:17:23.8934705Z Compile requests 1152 2025-12-04T10:17:23.8934886Z Compile requests executed 397 2025-12-04T10:17:23.8935012Z Cache hits 7 2025-12-04T10:17:23.8935134Z Cache hits (C/C++) 4 2025-12-04T10:17:23.8935254Z Cache hits (HIP) 3 2025-12-04T10:17:23.8935371Z Cache misses 390 2025-12-04T10:17:23.8935492Z Cache misses (C/C++) 373 2025-12-04T10:17:23.8935610Z Cache misses (HIP) 17 2025-12-04T10:17:23.8935739Z Cache hits rate 1.76 % 2025-12-04T10:17:23.8935865Z Cache hits rate (C/C++) 1.06 % 2025-12-04T10:17:23.8935992Z Cache hits rate (HIP) 15.00 % 2025-12-04T10:17:23.8936117Z Cache timeouts 0 2025-12-04T10:17:23.8936237Z Cache read errors 0 2025-12-04T10:17:23.8936355Z Forced recaches 0 2025-12-04T10:17:23.8936475Z Cache write errors 0 2025-12-04T10:17:23.8936593Z Cache errors 0 2025-12-04T10:17:23.8936710Z Compilations 390 2025-12-04T10:17:23.8936893Z Compilation failures 0 2025-12-04T10:17:23.8937015Z Non-cacheable compilations 0 2025-12-04T10:17:23.8937137Z Non-cacheable calls 152 2025-12-04T10:17:23.8937260Z Non-compilation calls 603 2025-12-04T10:17:23.8937388Z Unsupported compiler calls 0 2025-12-04T10:17:23.8937653Z Average cache write 0.000 s 2025-12-04T10:17:23.8937781Z Average compiler 3.560 s 2025-12-04T10:17:23.8937908Z Average cache read hit 0.000 s 2025-12-04T10:17:23.8938094Z Failed distributed compilations 0 2025-12-04T10:17:23.8938182Z 2025-12-04T10:17:23.8938227Z Non-cacheable reasons: 2025-12-04T10:17:23.8938338Z unknown source language 150 2025-12-04T10:17:23.8938461Z -E 2 2025-12-04T10:17:23.8938543Z 2025-12-04T10:17:23.8938633Z Cache location Local disk: "/var/lib/jenkins/.cache/sccache" 2025-12-04T10:17:23.8938800Z Use direct/preprocessor mode? yes 2025-12-04T10:17:23.8938928Z Version (client) 0.10.0 2025-12-04T10:17:23.8939055Z Cache size 46 MiB 2025-12-04T10:17:23.8939183Z Max cache size 10 GiB 2025-12-04T10:17:23.8939310Z + sccache --stop-server 2025-12-04T10:17:23.8950341Z Stopping sccache server... 2025-12-04T10:17:23.8952750Z Compile requests 1152 2025-12-04T10:17:23.8952906Z Compile requests executed 397 2025-12-04T10:17:23.8953032Z Cache hits 7 2025-12-04T10:17:23.8953156Z Cache hits (C/C++) 4 2025-12-04T10:17:23.8953348Z Cache hits (HIP) 3 2025-12-04T10:17:23.8953469Z Cache misses 390 2025-12-04T10:17:23.8953590Z Cache misses (C/C++) 373 2025-12-04T10:17:23.8953709Z Cache misses (HIP) 17 2025-12-04T10:17:23.8953833Z Cache hits rate 1.76 % 2025-12-04T10:17:23.8953960Z Cache hits rate (C/C++) 1.06 % 2025-12-04T10:17:23.8954088Z Cache hits rate (HIP) 15.00 % 2025-12-04T10:17:23.8954211Z Cache timeouts 0 2025-12-04T10:17:23.8954334Z Cache read errors 0 2025-12-04T10:17:23.8954454Z Forced recaches 0 2025-12-04T10:17:23.8954595Z Cache write errors 0 2025-12-04T10:17:23.8954720Z Cache errors 0 2025-12-04T10:17:23.8954841Z Compilations 390 2025-12-04T10:17:23.8954963Z Compilation failures 0 2025-12-04T10:17:23.8955092Z Non-cacheable compilations 0 2025-12-04T10:17:23.8955217Z Non-cacheable calls 152 2025-12-04T10:17:23.8955340Z Non-compilation calls 603 2025-12-04T10:17:23.8955468Z Unsupported compiler calls 0 2025-12-04T10:17:23.8955594Z Average cache write 0.000 s 2025-12-04T10:17:23.8955724Z Average compiler 3.560 s 2025-12-04T10:17:23.8955854Z Average cache read hit 0.000 s 2025-12-04T10:17:23.8955986Z Failed distributed compilations 0 2025-12-04T10:17:23.8956073Z 2025-12-04T10:17:23.8956117Z Non-cacheable reasons: 2025-12-04T10:17:23.8956229Z unknown source language 150 2025-12-04T10:17:23.8956356Z -E 2 2025-12-04T10:17:23.8956434Z 2025-12-04T10:17:23.8956519Z Cache location Local disk: "/var/lib/jenkins/.cache/sccache" 2025-12-04T10:17:23.8956689Z Use direct/preprocessor mode? yes 2025-12-04T10:17:23.8956860Z Version (client) 0.10.0 2025-12-04T10:17:23.8956990Z Cache size 46 MiB 2025-12-04T10:17:23.8957124Z Max cache size 10 GiB 2025-12-04T10:17:23.8957319Z + echo ::endgroup:: 2025-12-04T10:17:23.8957518Z ##[endgroup] 2025-12-04T10:17:23.9017158Z ##[group]Run # copy test results back to the mounted workspace, needed sudo, resulting permissions were correct 2025-12-04T10:17:23.9017502Z # copy test results back to the mounted workspace, needed sudo, resulting permissions were correct 2025-12-04T10:17:23.9017896Z docker exec -t "eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa" sh -c "cd ../pytorch && sudo cp -R test/test-reports ../workspace/test" 2025-12-04T10:17:23.9022326Z shell: /usr/bin/bash -e {0} 2025-12-04T10:17:23.9022456Z env: 2025-12-04T10:17:23.9022634Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:23.9022783Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:23.9022977Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:23.9023223Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:23.9023637Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:23.9024017Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:23.9024140Z AWS_REGION: us-east-1 2025-12-04T10:17:23.9024310Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:23.9024468Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:23.9026695Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:23.9026924Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:23.9027115Z ##[endgroup] 2025-12-04T10:17:23.9620098Z ##[group]Run docker exec -t "eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa" sh -c "sudo chown -R 1001:1001 test" 2025-12-04T10:17:23.9620512Z docker exec -t "eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa" sh -c "sudo chown -R 1001:1001 test" 2025-12-04T10:17:23.9625092Z shell: /usr/bin/bash -e {0} 2025-12-04T10:17:23.9625205Z env: 2025-12-04T10:17:23.9625299Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:23.9625437Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:23.9625614Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:23.9625780Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:23.9626162Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:23.9626546Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:23.9626674Z AWS_REGION: us-east-1 2025-12-04T10:17:23.9626905Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:23.9627057Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:23.9629246Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:23.9629418Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:23.9629600Z ##[endgroup] 2025-12-04T10:17:24.0359396Z ##[group]Run cat test/**/*_toprint.log || true 2025-12-04T10:17:24.0359570Z cat test/**/*_toprint.log || true 2025-12-04T10:17:24.0363671Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:17:24.0363820Z env: 2025-12-04T10:17:24.0363916Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:24.0364049Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:24.0364226Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:24.0364390Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:24.0364769Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:24.0365152Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:24.0365266Z AWS_REGION: us-east-1 2025-12-04T10:17:24.0365431Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:24.0365591Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:24.0367829Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:24.0368003Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:24.0368188Z ##[endgroup] 2025-12-04T10:17:24.0409944Z Test results will be stored in test-reports/python-pytest/test_ci_sanity_check_fail/test_ci_sanity_check_fail-57c359cd2c9207d9.xml 2025-12-04T10:17:24.0410269Z ============================= test session starts ============================== 2025-12-04T10:17:24.0410533Z platform linux -- Python 3.10.14, pytest-7.3.2, pluggy-1.6.0 -- /opt/conda/envs/py_3.10/bin/python 2025-12-04T10:17:24.0410979Z cachedir: .pytest_cache 2025-12-04T10:17:24.0411383Z hypothesis profile 'pytorch_ci' -> database=None, max_examples=50, derandomize=True, suppress_health_check=[HealthCheck.too_slow] 2025-12-04T10:17:24.0411742Z rootdir: /var/lib/jenkins/pytorch 2025-12-04T10:17:24.0411877Z configfile: pytest.ini 2025-12-04T10:17:24.0412151Z plugins: hypothesis-6.56.4, cpp-2.3.0, flakefinder-1.1.0, rerunfailures-14.0, subtests-0.13.1, xdist-3.3.1, xdoctest-1.3.0, typeguard-4.3.0 2025-12-04T10:17:24.0412427Z collecting ... collected 2 items 2025-12-04T10:17:24.0412587Z stepcurrent: Cannot find last run test, not skipping 2025-12-04T10:17:24.0412751Z Running 0 items in this shard: 2025-12-04T10:17:24.0412839Z 2025-12-04T10:17:24.0413084Z - generated xml file: /var/lib/jenkins/pytorch/test/test-reports/python-pytest/test_ci_sanity_check_fail/test_ci_sanity_check_fail-57c359cd2c9207d9.xml - 2025-12-04T10:17:24.0413449Z ============================ no tests ran in 0.01s ============================= 2025-12-04T10:17:24.0470436Z Prepare all required actions 2025-12-04T10:17:24.0470775Z Getting action download info 2025-12-04T10:17:24.6062252Z Download action repository 'seemethere/upload-artifact-s3@v5' (SHA:baba72d0712b404f646cebe0730933554ebce96a) 2025-12-04T10:17:25.6407820Z Download action repository 'actions/upload-artifact@v4' (SHA:ea165f8d65b6e75b540449e92b4886f43607fa02) 2025-12-04T10:17:27.0648027Z ##[group]Run ./.github/actions/upload-test-artifacts 2025-12-04T10:17:27.0648196Z with: 2025-12-04T10:17:27.0648293Z use-gha: true 2025-12-04T10:17:27.0648454Z file-suffix: test-default-4-6-linux.rocm.gpu.gfx942.1.b_57116213149 2025-12-04T10:17:27.0648633Z s3-bucket: gha-artifacts 2025-12-04T10:17:27.0648746Z env: 2025-12-04T10:17:27.0648845Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:27.0648986Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:27.0649171Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:27.0649382Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:27.0649793Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:27.0650182Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:27.0650305Z AWS_REGION: us-east-1 2025-12-04T10:17:27.0650472Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:27.0650629Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:27.0652848Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:27.0653025Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:27.0653213Z ##[endgroup] 2025-12-04T10:17:27.0683528Z ##[group]Run actions/upload-artifact@v4 2025-12-04T10:17:27.0683663Z with: 2025-12-04T10:17:27.0683847Z name: test-jsons-runattempt1-test-default-4-6-linux.rocm.gpu.gfx942.1.b_57116213149.zip 2025-12-04T10:17:27.0684056Z retention-days: 14 2025-12-04T10:17:27.0684175Z if-no-files-found: warn 2025-12-04T10:17:27.0684285Z path: test/**/*.json 2025-12-04T10:17:27.0684392Z compression-level: 6 2025-12-04T10:17:27.0684496Z overwrite: false 2025-12-04T10:17:27.0684604Z include-hidden-files: false 2025-12-04T10:17:27.0684715Z env: 2025-12-04T10:17:27.0684810Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:27.0684946Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:27.0685124Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:27.0685291Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:27.0685678Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:27.0686049Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:27.0686170Z AWS_REGION: us-east-1 2025-12-04T10:17:27.0686301Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:27.0686549Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:27.0688785Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:27.0688958Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:27.0689209Z ##[endgroup] 2025-12-04T10:17:27.4222444Z With the provided path, there will be 6 files uploaded 2025-12-04T10:17:27.4225157Z Artifact name is valid! 2025-12-04T10:17:27.4225891Z Root directory input is valid! 2025-12-04T10:17:27.7144480Z Beginning upload of artifact content to blob storage 2025-12-04T10:17:28.1236162Z Uploaded bytes 46621 2025-12-04T10:17:28.2369563Z Finished uploading artifact content to blob storage! 2025-12-04T10:17:28.2370870Z SHA256 digest of uploaded artifact zip is 43b6e14a6a9dbccb848a5ced9870e068794b12171d4c4e5e15c690985ec06ace 2025-12-04T10:17:28.2371635Z Finalizing artifact upload 2025-12-04T10:17:28.3982436Z Artifact test-jsons-runattempt1-test-default-4-6-linux.rocm.gpu.gfx942.1.b_57116213149.zip.zip successfully finalized. Artifact ID 4762517602 2025-12-04T10:17:28.3983649Z Artifact test-jsons-runattempt1-test-default-4-6-linux.rocm.gpu.gfx942.1.b_57116213149.zip has been successfully uploaded! Final size is 46621 bytes. Artifact ID is 4762517602 2025-12-04T10:17:28.3987802Z Artifact download URL: https://github.com/pytorch/pytorch/actions/runs/19922849170/artifacts/4762517602 2025-12-04T10:17:28.4105941Z ##[group]Run actions/upload-artifact@v4 2025-12-04T10:17:28.4106089Z with: 2025-12-04T10:17:28.4106286Z name: test-reports-runattempt1-test-default-4-6-linux.rocm.gpu.gfx942.1.b_57116213149.zip 2025-12-04T10:17:28.4106504Z retention-days: 14 2025-12-04T10:17:28.4106623Z if-no-files-found: ignore 2025-12-04T10:17:28.4106805Z path: test/**/*.xml test/**/*.csv 2025-12-04T10:17:28.4106932Z compression-level: 6 2025-12-04T10:17:28.4107041Z overwrite: false 2025-12-04T10:17:28.4107151Z include-hidden-files: false 2025-12-04T10:17:28.4107265Z env: 2025-12-04T10:17:28.4107364Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:28.4107529Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:28.4107716Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:28.4107899Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:28.4108316Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:28.4108700Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:28.4108823Z AWS_REGION: us-east-1 2025-12-04T10:17:28.4109019Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:28.4109179Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:28.4111359Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:28.4111536Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:28.4111726Z ##[endgroup] 2025-12-04T10:17:28.8155540Z With the provided path, there will be 99 files uploaded 2025-12-04T10:17:28.8158598Z Artifact name is valid! 2025-12-04T10:17:28.8159185Z Root directory input is valid! 2025-12-04T10:17:29.0911634Z Beginning upload of artifact content to blob storage 2025-12-04T10:17:29.8474308Z Uploaded bytes 233660 2025-12-04T10:17:29.9343363Z Finished uploading artifact content to blob storage! 2025-12-04T10:17:29.9344086Z SHA256 digest of uploaded artifact zip is 77d1c1511cb1e06da5531f7b5e3217305c3b4b2a2b24d983c4049d1e30557b0f 2025-12-04T10:17:29.9345009Z Finalizing artifact upload 2025-12-04T10:17:30.1018048Z Artifact test-reports-runattempt1-test-default-4-6-linux.rocm.gpu.gfx942.1.b_57116213149.zip.zip successfully finalized. Artifact ID 4762517913 2025-12-04T10:17:30.1018612Z Artifact test-reports-runattempt1-test-default-4-6-linux.rocm.gpu.gfx942.1.b_57116213149.zip has been successfully uploaded! Final size is 233660 bytes. Artifact ID is 4762517913 2025-12-04T10:17:30.1022856Z Artifact download URL: https://github.com/pytorch/pytorch/actions/runs/19922849170/artifacts/4762517913 2025-12-04T10:17:30.1135524Z ##[group]Run actions/upload-artifact@v4 2025-12-04T10:17:30.1135692Z with: 2025-12-04T10:17:30.1135874Z name: logs-runattempt1-test-default-4-6-linux.rocm.gpu.gfx942.1.b_57116213149.zip 2025-12-04T10:17:30.1136155Z retention-days: 14 2025-12-04T10:17:30.1136274Z if-no-files-found: ignore 2025-12-04T10:17:30.1136401Z path: usage_log.txt test/**/*.log 2025-12-04T10:17:30.1136529Z compression-level: 6 2025-12-04T10:17:30.1136639Z overwrite: false 2025-12-04T10:17:30.1136809Z include-hidden-files: false 2025-12-04T10:17:30.1136929Z env: 2025-12-04T10:17:30.1137025Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:30.1137169Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:30.1137352Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:30.1137525Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:30.1138037Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:30.1138424Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:30.1138548Z AWS_REGION: us-east-1 2025-12-04T10:17:30.1138722Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:30.1138882Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:30.1141082Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:30.1141265Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:30.1141456Z ##[endgroup] 2025-12-04T10:17:30.5312028Z Multiple search paths detected. Calculating the least common ancestor of all paths 2025-12-04T10:17:30.5313095Z The least common ancestor is /home/runner/_work/pytorch/pytorch. This will be the root directory of the artifact 2025-12-04T10:17:30.5313369Z With the provided path, there will be 97 files uploaded 2025-12-04T10:17:30.5316375Z Artifact name is valid! 2025-12-04T10:17:30.5317791Z Root directory input is valid! 2025-12-04T10:17:30.8200895Z Beginning upload of artifact content to blob storage 2025-12-04T10:17:31.3745638Z Uploaded bytes 156699 2025-12-04T10:17:31.4530317Z Finished uploading artifact content to blob storage! 2025-12-04T10:17:31.4531931Z SHA256 digest of uploaded artifact zip is 2ee6c4977fd9f87d499a6de28d7dc71180a75c612f9377adcbb785a2df74b2d9 2025-12-04T10:17:31.4532714Z Finalizing artifact upload 2025-12-04T10:17:31.7100505Z Artifact logs-runattempt1-test-default-4-6-linux.rocm.gpu.gfx942.1.b_57116213149.zip.zip successfully finalized. Artifact ID 4762518178 2025-12-04T10:17:31.7101369Z Artifact logs-runattempt1-test-default-4-6-linux.rocm.gpu.gfx942.1.b_57116213149.zip has been successfully uploaded! Final size is 156699 bytes. Artifact ID is 4762518178 2025-12-04T10:17:31.7105900Z Artifact download URL: https://github.com/pytorch/pytorch/actions/runs/19922849170/artifacts/4762518178 2025-12-04T10:17:31.7255783Z ##[group]Run # shellcheck disable=SC2156 2025-12-04T10:17:31.7255987Z # shellcheck disable=SC2156 2025-12-04T10:17:31.7256233Z find . -iname "core.[1-9]*" -exec docker exec "${CONTAINER_NAME}" sh -c "gdb python {} -ex 'bt' -ex 'q'" \; 2025-12-04T10:17:31.7260477Z shell: /usr/bin/bash -e {0} 2025-12-04T10:17:31.7260611Z env: 2025-12-04T10:17:31.7260717Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:31.7260874Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:31.7261070Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:31.7261250Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:31.7261649Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:31.7262040Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:31.7262190Z AWS_REGION: us-east-1 2025-12-04T10:17:31.7262389Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:31.7262654Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:31.7264907Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:31.7265091Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:31.7265363Z ##[endgroup] 2025-12-04T10:17:31.8610413Z ##[group]Run aws-actions/configure-aws-credentials@ececac1a45f3b08a01d2dd070d28d111c5fe6722 2025-12-04T10:17:31.8610639Z with: 2025-12-04T10:17:31.8610808Z role-to-assume: arn:aws:iam::308535385114:role/gha_workflow_upload-benchmark-results 2025-12-04T10:17:31.8611009Z role-duration-seconds: 18000 2025-12-04T10:17:31.8611136Z aws-region: us-east-1 2025-12-04T10:17:31.8611262Z audience: sts.amazonaws.com 2025-12-04T10:17:31.8611385Z env: 2025-12-04T10:17:31.8611486Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:31.8611631Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:31.8611821Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:31.8612002Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:31.8612426Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:31.8612816Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:31.8612942Z AWS_REGION: us-east-1 2025-12-04T10:17:31.8613115Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:31.8613282Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:31.8615493Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:31.8615677Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:31.8615866Z ##[endgroup] 2025-12-04T10:17:32.1861240Z Assuming role with OIDC 2025-12-04T10:17:32.5479643Z Authenticated as assumedRoleId AROAUPVRELQNA5GQHA6IA:GitHubActions 2025-12-04T10:17:32.6486146Z ##[group]Run pytorch/test-infra/.github/actions/upload-benchmark-results@main 2025-12-04T10:17:32.6486369Z with: 2025-12-04T10:17:32.6486514Z benchmark-results-dir: test/test-reports 2025-12-04T10:17:32.6486669Z dry-run: false 2025-12-04T10:17:32.6487065Z schema-version: v3 2025-12-04T10:17:32.6487362Z github-token: *** 2025-12-04T10:17:32.6487485Z env: 2025-12-04T10:17:32.6487599Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:32.6487757Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:32.6487958Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:32.6488145Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:32.6488547Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:32.6488956Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:32.6489286Z AWS_REGION: us-east-1 2025-12-04T10:17:32.6489432Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:32.6489613Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:32.6491897Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:32.6492088Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:32.6492286Z ##[endgroup] 2025-12-04T10:17:32.6501059Z ##[group]Run set -eux 2025-12-04T10:17:32.6501214Z set -eux 2025-12-04T10:17:32.6501329Z  2025-12-04T10:17:32.6501436Z if [[ -n "" ]]; then 2025-12-04T10:17:32.6501572Z  source "" 2025-12-04T10:17:32.6501690Z fi 2025-12-04T10:17:32.6501849Z python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-12-04T10:17:32.6502030Z  2025-12-04T10:17:32.6502141Z DEVICE_NAME="" 2025-12-04T10:17:32.6502265Z DEVICE_TYPE="" 2025-12-04T10:17:32.6502388Z  2025-12-04T10:17:32.6502508Z if command -v nvidia-smi; then 2025-12-04T10:17:32.6502705Z  # NB: I'm using PyTorch here to get the device name, however, it needs to 2025-12-04T10:17:32.6503041Z  # install the correct version of PyTorch manually for now. Any PyTorch 2025-12-04T10:17:32.6503270Z  # version is fine, I just use 2.7.1 to satify PYPIDEP linter 2025-12-04T10:17:32.6503538Z  python3 -mpip install torch==2.7.1 2025-12-04T10:17:32.6503697Z elif command -v rocminfo; then 2025-12-04T10:17:32.6503892Z  # NB: Installing torch on ROCm runner with pip here causes CI to fail 2025-12-04T10:17:32.6504122Z  # with a memoryview is too large error only on MI300 runners. Is pip 2025-12-04T10:17:32.6504350Z  # version on ROCm runner there too old? As a workaround, let's use the 2025-12-04T10:17:32.6504563Z  # GPU device name coming from rocminfo instead 2025-12-04T10:17:32.6504731Z  DEVICE_NAME=rocm 2025-12-04T10:17:32.6504946Z  DEVICE_TYPE=$(rocminfo | grep "Marketing Name" | tail -n1 | awk -F':' '{print $2}' | xargs) 2025-12-04T10:17:32.6505151Z fi 2025-12-04T10:17:32.6505266Z  2025-12-04T10:17:32.6505404Z echo "DEVICE_NAME=$DEVICE_NAME" >> $GITHUB_ENV 2025-12-04T10:17:32.6505584Z echo "DEVICE_TYPE=$DEVICE_TYPE" >> $GITHUB_ENV 2025-12-04T10:17:32.6509722Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:17:32.6509896Z env: 2025-12-04T10:17:32.6510012Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:32.6510178Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:32.6510383Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:32.6510575Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:32.6511056Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:32.6511453Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:32.6511589Z AWS_REGION: us-east-1 2025-12-04T10:17:32.6511747Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:32.6511928Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:32.6514233Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:32.6514434Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:32.6514639Z ##[endgroup] 2025-12-04T10:17:32.6534584Z + [[ -n '' ]] 2025-12-04T10:17:32.6534952Z + python3 -mpip install boto3==1.35.33 psutil==7.0.0 pynvml==12.0.0 2025-12-04T10:17:32.7948569Z Defaulting to user installation because normal site-packages is not writeable 2025-12-04T10:17:33.1527830Z Collecting boto3==1.35.33 2025-12-04T10:17:33.1892916Z Downloading boto3-1.35.33-py3-none-any.whl (139 kB) 2025-12-04T10:17:33.2068473Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 139.1/139.1 KB 8.7 MB/s eta 0:00:00 2025-12-04T10:17:33.2834767Z Collecting psutil==7.0.0 2025-12-04T10:17:33.2900182Z Downloading psutil-7.0.0-cp36-abi3-manylinux_2_12_x86_64.manylinux2010_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (277 kB) 2025-12-04T10:17:33.3072856Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 278.0/278.0 KB 16.6 MB/s eta 0:00:00 2025-12-04T10:17:33.3190217Z Collecting pynvml==12.0.0 2025-12-04T10:17:33.3254483Z Downloading pynvml-12.0.0-py3-none-any.whl (26 kB) 2025-12-04T10:17:33.3435945Z Collecting s3transfer<0.11.0,>=0.10.0 2025-12-04T10:17:33.3487890Z Downloading s3transfer-0.10.4-py3-none-any.whl (83 kB) 2025-12-04T10:17:33.3536272Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 83.2/83.2 KB 18.7 MB/s eta 0:00:00 2025-12-04T10:17:33.7307986Z Collecting botocore<1.36.0,>=1.35.33 2025-12-04T10:17:33.7887662Z Downloading botocore-1.35.99-py3-none-any.whl (13.3 MB) 2025-12-04T10:17:34.0509273Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 13.3/13.3 MB 61.8 MB/s eta 0:00:00 2025-12-04T10:17:34.0766540Z Collecting jmespath<2.0.0,>=0.7.1 2025-12-04T10:17:34.0820623Z Downloading jmespath-1.0.1-py3-none-any.whl (20 kB) 2025-12-04T10:17:34.1046534Z Collecting nvidia-ml-py<13.0.0a0,>=12.0.0 2025-12-04T10:17:34.1106218Z Downloading nvidia_ml_py-12.575.51-py3-none-any.whl (47 kB) 2025-12-04T10:17:34.1125668Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 47.5/47.5 KB 65.0 MB/s eta 0:00:00 2025-12-04T10:17:34.1286573Z Collecting python-dateutil<3.0.0,>=2.1 2025-12-04T10:17:34.1363188Z Downloading python_dateutil-2.9.0.post0-py2.py3-none-any.whl (229 kB) 2025-12-04T10:17:34.1390818Z ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 229.9/229.9 KB 137.4 MB/s eta 0:00:00 2025-12-04T10:17:34.1402859Z Requirement already satisfied: urllib3!=2.2.0,<3,>=1.25.4 in /home/runner/.local/lib/python3.10/site-packages (from botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.26.20) 2025-12-04T10:17:34.2164749Z Requirement already satisfied: six>=1.5 in /usr/lib/python3/dist-packages (from python-dateutil<3.0.0,>=2.1->botocore<1.36.0,>=1.35.33->boto3==1.35.33) (1.16.0) 2025-12-04T10:17:34.2873985Z Installing collected packages: nvidia-ml-py, python-dateutil, pynvml, psutil, jmespath, botocore, s3transfer, boto3 2025-12-04T10:17:34.6266153Z Successfully installed boto3-1.35.33 botocore-1.35.99 jmespath-1.0.1 nvidia-ml-py-12.575.51 psutil-7.0.0 pynvml-12.0.0 python-dateutil-2.9.0.post0 s3transfer-0.10.4 2025-12-04T10:17:34.7191651Z + DEVICE_NAME= 2025-12-04T10:17:34.7192010Z + DEVICE_TYPE= 2025-12-04T10:17:34.7192380Z + command -v nvidia-smi 2025-12-04T10:17:34.7192733Z + command -v rocminfo 2025-12-04T10:17:34.7193277Z /usr/bin/rocminfo 2025-12-04T10:17:34.7193597Z + DEVICE_NAME=rocm 2025-12-04T10:17:34.7201523Z ++ rocminfo 2025-12-04T10:17:34.7201720Z ++ grep 'Marketing Name' 2025-12-04T10:17:34.7203249Z ++ tail -n1 2025-12-04T10:17:34.7204582Z ++ awk -F: '{print $2}' 2025-12-04T10:17:34.7205505Z ++ xargs 2025-12-04T10:17:34.7755047Z + DEVICE_TYPE='AMD Instinct MI325X' 2025-12-04T10:17:34.7755476Z + echo DEVICE_NAME=rocm 2025-12-04T10:17:34.7755850Z + echo 'DEVICE_TYPE=AMD Instinct MI325X' 2025-12-04T10:17:34.7778900Z ##[group]Run set -eux 2025-12-04T10:17:34.7779079Z set -eux 2025-12-04T10:17:34.7779203Z  2025-12-04T10:17:34.7779341Z if [[ -z "${GITHUB_TOKEN}" ]]; then 2025-12-04T10:17:34.7779532Z  echo "Missing github-token input" 2025-12-04T10:17:34.7779695Z  exit 1 2025-12-04T10:17:34.7779818Z fi 2025-12-04T10:17:34.7784446Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:17:34.7784634Z env: 2025-12-04T10:17:34.7784756Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:34.7784934Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:34.7785156Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:34.7785367Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:34.7785828Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:34.7786281Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:34.7786432Z AWS_REGION: us-east-1 2025-12-04T10:17:34.7786638Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:34.7786886Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:34.7789414Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:34.7789610Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:34.7789810Z DEVICE_NAME: rocm 2025-12-04T10:17:34.7789930Z DEVICE_TYPE: AMD Instinct MI325X 2025-12-04T10:17:34.7790168Z GITHUB_TOKEN: *** 2025-12-04T10:17:34.7790282Z ##[endgroup] 2025-12-04T10:17:34.7810701Z + [[ -z *** ]] 2025-12-04T10:17:34.7843126Z ##[group]Run pytorch/test-infra/.github/actions/get-workflow-job-id@main 2025-12-04T10:17:34.7843327Z with: 2025-12-04T10:17:34.7843498Z github-token: *** 2025-12-04T10:17:34.7843612Z env: 2025-12-04T10:17:34.7843718Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:34.7843888Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:34.7844089Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:34.7844278Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:34.7844760Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:34.7845212Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:34.7845348Z AWS_REGION: us-east-1 2025-12-04T10:17:34.7845487Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:34.7845661Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:34.7848170Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:34.7848360Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:34.7848557Z DEVICE_NAME: rocm 2025-12-04T10:17:34.7848681Z DEVICE_TYPE: AMD Instinct MI325X 2025-12-04T10:17:34.7848818Z ##[endgroup] 2025-12-04T10:17:34.7855121Z ##[group]Run set -eux 2025-12-04T10:17:34.7855263Z set -eux 2025-12-04T10:17:34.7855371Z  2025-12-04T10:17:34.7855571Z python3 "${GITHUB_ACTION_PATH}/../../scripts/get_workflow_job_id.py" "${GITHUB_RUN_ID}" "${RUNNER_NAME}" 2025-12-04T10:17:34.7860123Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:17:34.7860288Z env: 2025-12-04T10:17:34.7860396Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:34.7860548Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:34.7860747Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:34.7860933Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:34.7861332Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:34.7861715Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:34.7861973Z AWS_REGION: us-east-1 2025-12-04T10:17:34.7862120Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:34.7862294Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:34.7864583Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:34.7864770Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:34.7864970Z DEVICE_NAME: rocm 2025-12-04T10:17:34.7865092Z DEVICE_TYPE: AMD Instinct MI325X 2025-12-04T10:17:34.7865278Z GITHUB_TOKEN: *** 2025-12-04T10:17:34.7865390Z ##[endgroup] 2025-12-04T10:17:34.7883900Z + python3 /home/runner/_work/_actions/pytorch/test-infra/main/.github/actions/get-workflow-job-id/../../scripts/get_workflow_job_id.py 19922849170 linux.rocm.gpu.gfx942.1.b-gwk9b-runner-68hm6 2025-12-04T10:17:36.1683833Z setting job-id=57116213149 2025-12-04T10:17:36.1684690Z setting job-name=linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable) 2025-12-04T10:17:36.1817531Z ##[group]Run set -eux 2025-12-04T10:17:36.1817737Z set -eux 2025-12-04T10:17:36.1817871Z  2025-12-04T10:17:36.1818013Z if [[ -n "" ]]; then 2025-12-04T10:17:36.1818195Z  source "" 2025-12-04T10:17:36.1818336Z fi 2025-12-04T10:17:36.1818472Z  2025-12-04T10:17:36.1818684Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_metadata.py" \ 2025-12-04T10:17:36.1818968Z  --schema-version "${SCHEMA_VERSION}" \ 2025-12-04T10:17:36.1819155Z  --repo "${REPO}" \ 2025-12-04T10:17:36.1819340Z  --head-branch "${HEAD_BRANCH}" \ 2025-12-04T10:17:36.1819521Z  --head-sha "${HEAD_SHA}" \ 2025-12-04T10:17:36.1819713Z  --workflow-id "${WORKFLOW_RUN_ID}" \ 2025-12-04T10:17:36.1819911Z  --run-attempt "${RUN_ATTEMPT}" \ 2025-12-04T10:17:36.1820099Z  --job-id "${JOB_ID}" \ 2025-12-04T10:17:36.1820280Z  --job-name "${JOB_NAME}" 2025-12-04T10:17:36.1824722Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:17:36.1824891Z env: 2025-12-04T10:17:36.1825012Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:36.1825178Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:36.1825456Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:36.1825652Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:36.1826128Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:36.1826525Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:36.1826669Z AWS_REGION: us-east-1 2025-12-04T10:17:36.1827095Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:36.1827283Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:36.1829565Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:36.1829771Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:36.1829975Z DEVICE_NAME: rocm 2025-12-04T10:17:36.1830113Z DEVICE_TYPE: AMD Instinct MI325X 2025-12-04T10:17:36.1830261Z SCHEMA_VERSION: v3 2025-12-04T10:17:36.1830389Z REPO: pytorch/pytorch 2025-12-04T10:17:36.1830524Z HEAD_BRANCH: refs/heads/main 2025-12-04T10:17:36.1830685Z HEAD_SHA: ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 2025-12-04T10:17:36.1830853Z WORKFLOW_RUN_ID: 19922849170 2025-12-04T10:17:36.1830990Z RUN_ATTEMPT: 1 2025-12-04T10:17:36.1831109Z JOB_ID: 57116213149 2025-12-04T10:17:36.1831339Z JOB_NAME: linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable) 2025-12-04T10:17:36.1831578Z ##[endgroup] 2025-12-04T10:17:36.1853322Z + [[ -n '' ]] 2025-12-04T10:17:36.1855222Z + python3 /home/runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_metadata.py --schema-version v3 --repo pytorch/pytorch --head-branch refs/heads/main --head-sha ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32 --workflow-id 19922849170 --run-attempt 1 --job-id 57116213149 --job-name 'linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable)' 2025-12-04T10:17:36.2032364Z ##[group]Run set -eux 2025-12-04T10:17:36.2032519Z set -eux 2025-12-04T10:17:36.2032639Z  2025-12-04T10:17:36.2032755Z if [[ -n "" ]]; then 2025-12-04T10:17:36.2032887Z  source "" 2025-12-04T10:17:36.2033009Z fi 2025-12-04T10:17:36.2033116Z  2025-12-04T10:17:36.2033296Z python3 "${GITHUB_ACTION_PATH}/../../scripts/benchmarks/gather_runners_info.py" 2025-12-04T10:17:36.2037856Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:17:36.2038030Z env: 2025-12-04T10:17:36.2038144Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:36.2038307Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:36.2038517Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:36.2038706Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:36.2039123Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:36.2039540Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:36.2039677Z AWS_REGION: us-east-1 2025-12-04T10:17:36.2039864Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:36.2040037Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:36.2042500Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:36.2042697Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:36.2042893Z DEVICE_NAME: rocm 2025-12-04T10:17:36.2043017Z DEVICE_TYPE: AMD Instinct MI325X 2025-12-04T10:17:36.2043155Z ##[endgroup] 2025-12-04T10:17:36.2060322Z + [[ -n '' ]] 2025-12-04T10:17:36.2060636Z + python3 /home/runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/benchmarks/gather_runners_info.py 2025-12-04T10:17:36.2270288Z INFO:root:Fail to import torch to get the device name 2025-12-04T10:17:36.2357233Z ##[group]Run set -eux 2025-12-04T10:17:36.2357412Z set -eux 2025-12-04T10:17:36.2357550Z  2025-12-04T10:17:36.2357804Z # TODO (huydhn): Implement this part 2025-12-04T10:17:36.2358021Z echo "dependencies={}" >> "${GITHUB_OUTPUT}" 2025-12-04T10:17:36.2362790Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:17:36.2363001Z env: 2025-12-04T10:17:36.2363148Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:36.2363350Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:36.2363595Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:36.2363820Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:36.2364333Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:36.2364835Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:36.2364997Z AWS_REGION: us-east-1 2025-12-04T10:17:36.2365267Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:36.2365497Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:36.2368062Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:36.2368259Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:36.2368459Z DEVICE_NAME: rocm 2025-12-04T10:17:36.2368586Z DEVICE_TYPE: AMD Instinct MI325X 2025-12-04T10:17:36.2368730Z ##[endgroup] 2025-12-04T10:17:36.2394501Z + echo 'dependencies={}' 2025-12-04T10:17:36.2417336Z ##[group]Run set -eux 2025-12-04T10:17:36.2417550Z set -eux 2025-12-04T10:17:36.2417687Z  2025-12-04T10:17:36.2417819Z if [[ -n "" ]]; then 2025-12-04T10:17:36.2417970Z  source "" 2025-12-04T10:17:36.2418252Z fi 2025-12-04T10:17:36.2418383Z  2025-12-04T10:17:36.2418529Z if [[ ! -d "${BENCHMARK_RESULTS_DIR}" ]]; then 2025-12-04T10:17:36.2418758Z  echo "${BENCHMARK_RESULTS_DIR} does not exist, skipping" 2025-12-04T10:17:36.2418994Z  # We don't want the job to fail if the directory doesn't exist 2025-12-04T10:17:36.2419193Z  exit 0 2025-12-04T10:17:36.2419320Z fi 2025-12-04T10:17:36.2419438Z  2025-12-04T10:17:36.2419567Z if [[ "${DRY_RUN}" == "true" ]]; then 2025-12-04T10:17:36.2419804Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-12-04T10:17:36.2420071Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-12-04T10:17:36.2420281Z  --metadata "${BENCHMARK_METADATA}" \ 2025-12-04T10:17:36.2420455Z  --runners "${RUNNER_INFO}" \ 2025-12-04T10:17:36.2420636Z  --dependencies "${DEPENDENCIES}" \ 2025-12-04T10:17:36.2420810Z  --dry-run 2025-12-04T10:17:36.2420944Z else 2025-12-04T10:17:36.2421139Z  python3 "${GITHUB_ACTION_PATH}/../../scripts/upload_benchmark_results.py" \ 2025-12-04T10:17:36.2421403Z  --benchmark-results-dir "${BENCHMARK_RESULTS_DIR}" \ 2025-12-04T10:17:36.2421609Z  --metadata "${BENCHMARK_METADATA}" \ 2025-12-04T10:17:36.2421794Z  --runners "${RUNNER_INFO}" \ 2025-12-04T10:17:36.2421974Z  --dependencies "${DEPENDENCIES}" 2025-12-04T10:17:36.2422133Z fi 2025-12-04T10:17:36.2426672Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:17:36.2427070Z env: 2025-12-04T10:17:36.2427199Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:36.2427366Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:36.2427572Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:36.2427761Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:36.2428181Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:36.2428654Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:36.2428859Z AWS_REGION: us-east-1 2025-12-04T10:17:36.2429073Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:36.2429259Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:36.2431562Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:36.2431768Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:36.2431976Z DEVICE_NAME: rocm 2025-12-04T10:17:36.2432116Z DEVICE_TYPE: AMD Instinct MI325X 2025-12-04T10:17:36.2432282Z BENCHMARK_RESULTS_DIR: test/test-reports 2025-12-04T10:17:36.2432428Z DRY_RUN: false 2025-12-04T10:17:36.2432978Z BENCHMARK_METADATA: {"timestamp": 1764843456, "schema_version": "v3", "name": "linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32", "workflow_id": 19922849170, "run_attempt": 1, "job_id": 57116213149} 2025-12-04T10:17:36.2433715Z RUNNER_INFO: [{"cpu_info": "x86_64", "cpu_count": 128, "avail_mem_in_gb": 3023, "extra_info": {"hostname": "linux.rocm.gpu.gfx942.1.b-gwk9b-runner-68hm6"}, "name": "rocm", "type": "AMD Instinct MI325X"}] 2025-12-04T10:17:36.2434041Z DEPENDENCIES: {} 2025-12-04T10:17:36.2434166Z ##[endgroup] 2025-12-04T10:17:36.2451678Z + [[ -n '' ]] 2025-12-04T10:17:36.2451843Z + [[ ! -d test/test-reports ]] 2025-12-04T10:17:36.2452011Z + [[ false == \t\r\u\e ]] 2025-12-04T10:17:36.2453457Z + python3 /home/runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/upload_benchmark_results.py --benchmark-results-dir test/test-reports --metadata '{"timestamp": 1764843456, "schema_version": "v3", "name": "linux-jammy-rocm-py3.10 / test (default, 4, 6, linux.rocm.gpu.gfx942.1.b, rerun_disabled_tests, unstable)", "repo": "pytorch/pytorch", "head_branch": "refs/heads/main", "head_sha": "ffd9b0fb4355e97af82fc42cf185c3ffa0fc0a32", "workflow_id": 19922849170, "run_attempt": 1, "job_id": 57116213149}' --runners '[{"cpu_info": "x86_64", "cpu_count": 128, "avail_mem_in_gb": 3023, "extra_info": {"hostname": "linux.rocm.gpu.gfx942.1.b-gwk9b-runner-68hm6"}, "name": "rocm", "type": "AMD Instinct MI325X"}]' --dependencies '{}' 2025-12-04T10:17:36.3103341Z /home/runner/_work/_actions/pytorch/test-infra/main/.github/actions/upload-benchmark-results/../../scripts/upload_benchmark_results.py:236: UserWarning: {'included': [{'test_file': 'inductor/test_collective_autotuning'}, {'test_file': 'test_public_bindings'}, {'test_file': 'inductor/test_aot_inductor'}, {'test_file': 'inductor/test_torchinductor'}, {'test_file': 'inductor/test_select_algorithm'}, {'test_file': 'inductor/test_torchinductor_dynamic_shapes'}, {'test_file': 'inductor/test_torchinductor_codegen_dynamic_shapes'}, {'test_file': 'inductor/test_kernel_benchmark'}, {'test_file': 'inductor/test_torchinductor_opinfo'}, {'test_file': 'inductor/test_pattern_matcher'}, {'test_file': 'inductor/test_group_batch_fusion'}, {'test_file': 'dynamo/test_dynamic_shapes'}, {'test_file': 'inductor/test_cpu_repro'}, {'test_file': 'inductor/test_cuda_repro'}, {'test_file': 'inductor/test_smoke'}, {'test_file': 'dynamo/test_after_aot'}, {'test_file': 'inductor/test_snode_runtime'}, {'test_file': 'inductor/test_minifier'}, {'test_file': 'inductor/test_compiled_autograd'}, {'test_file': 'inductor/test_custom_lowering'}, {'test_file': 'inductor/test_cudagraph_trees'}, {'test_file': 'inductor/test_perf'}, {'test_file': 'inductor/test_fused_attention'}, {'test_file': 'dynamo/test_activation_checkpointing'}, {'test_file': 'inductor/test_binary_folding'}, {'test_file': 'inductor/test_mkldnn_pattern_matcher'}, {'test_file': 'dynamo/test_logging'}, {'test_file': 'inductor/test_inductor_freezing'}, {'test_file': 'inductor/test_layout_optim'}, {'test_file': 'dynamo/test_unspec'}, {'test_file': 'dynamo/test_repros'}, {'test_file': 'dynamo/test_higher_order_ops'}, {'test_file': 'inductor/test_mmdecomp'}, {'test_file': 'dynamo/test_ctx_manager'}, {'test_file': 'dynamo/test_exc'}, {'test_file': 'dynamo/test_misc'}, {'test_file': 'inductor/test_cpu_select_algorithm'}, {'test_file': 'inductor/test_flex_attention'}, {'test_file': 'inductor/test_cuda_select_algorithm'}, {'test_file': 'inductor/test_aot_inductor_arrayref'}, {'test_file': 'inductor/test_cpu_cpp_wrapper'}, {'test_file': 'inductor/test_cutlass_backend'}, {'test_file': 'inductor/test_triton_cpu_backend'}, {'test_file': 'inductor/test_torchinductor_strided_blocks'}, {'test_file': 'test_custom_ops'}, {'test_file': 'test_content_store'}, {'test_file': 'inductor/test_halide'}, {'test_file': 'inductor/test_compile_subprocess'}, {'test_file': 'inductor/test_flex_decoding'}, {'test_file': 'inductor/test_deterministic'}, {'test_file': 'inductor/test_multi_kernel'}, {'test_file': 'inductor/test_analysis'}, {'test_file': 'inductor/test_pad_mm'}, {'test_file': 'inductor/test_triton_syntax'}, {'test_file': 'inductor/test_triton_extension_backend'}, {'test_file': 'test_sparse_semi_structured'}, {'test_file': 'inductor/test_op_completeness'}, {'test_file': 'inductor/test_subgraph_choice'}, {'test_file': 'inductor/test_b2b_gemm'}, {'test_file': 'inductor/test_triton_heuristics'}, {'test_file': 'inductor/test_cutedsl_grouped_mm'}, {'test_file': 'inductor/test_cpp_wrapper_hipify'}, {'test_file': 'export/test_retraceability'}, {'test_file': 'inductor/test_ck_backend'}, {'test_file': 'inductor/test_inductor_utils'}, {'test_file': 'inductor/test_template_heuristics_registry'}, {'test_file': 'inductor/test_async_compile'}, {'test_file': 'inductor/test_gpu_cpp_wrapper'}, {'test_file': 'export/test_export_training_ir_to_run_decomp'}, {'test_file': 'dynamo/test_deque_reconstruct'}, {'test_file': 'inductor/test_utils'}, {'test_file': 'inductor/test_indexing'}, {'test_file': 'inductor/test_inductor_annotations'}, {'test_file': 'inductor/test_compile_worker'}, {'test_file': 'dynamo/test_einops'}, {'test_file': 'inductor/test_external_callables'}, {'test_file': 'test_testing'}, {'test_file': 'dynamo/test_fx_passes_pre_grad'}, {'test_file': 'inductor/test_fp8'}, {'test_file': 'inductor/test_autoheuristic'}, {'test_file': 'export/test_strict_export_v2'}, {'test_file': 'inductor/test_flex_flash'}, {'test_file': 'dynamo/test_model_output'}, {'test_file': 'inductor/test_segmented_tree'}, {'test_file': 'inductor/test_kernel_optimization'}, {'test_file': 'inductor/test_metrics'}, {'test_file': 'export/test_unflatten_training_ir'}, {'test_file': 'inductor/test_triton_kernels'}, {'test_file': 'inductor/test_fx_fusion'}, {'test_file': 'inductor/test_extension_backend'}, {'test_file': 'inductor/test_native_matmul'}, {'test_file': 'inductor/test_xpu_basic'}, {'test_file': 'dynamo/test_inline_and_install'}, {'test_file': 'export/test_functionalized_assertions'}, {'test_file': 'inductor/test_loop_ordering'}, {'test_file': 'inductor/test_selective_lowering'}, {'test_file': 'dynamo/test_base_output'}, {'test_file': 'inductor/test_lookup_table'}, {'test_file': 'inductor/test_cooperative_reductions'}, {'test_file': 'export/test_serialize'}, {'test_file': 'inductor/test_cutedsl_template'}, {'test_file': 'inductor/test_benchmark_fusion'}, {'test_file': 'inductor/test_inductor_scheduler'}, {'test_file': 'inductor/test_move_constructors_to_gpu'}, {'test_file': 'export/test_export_strict'}, {'test_file': 'dynamo/test_modules'}, {'test_file': 'export/test_serdes'}, {'test_file': 'dynamo/test_regional_inductor'}, {'test_file': 'inductor/test_remote_cache'}, {'test_file': 'inductor/test_coordinate_descent_tuner'}, {'test_file': 'inductor/test_inplace_padding'}, {'test_file': 'inductor/test_cudacodecache'}, {'test_file': 'inductor/test_minifier_utils'}, {'test_file': 'inductor/test_debug_trace'}, {'test_file': 'dynamo/test_fx_graph_runnable'}, {'test_file': 'dynamo/test_recompiles'}, {'test_file': 'inductor/test_foreach'}, {'test_file': 'export/test_tree_utils'}, {'test_file': 'dynamo/test_backends'}, {'test_file': 'inductor/test_triton_wrapper'}, {'test_file': 'inductor/test_static_cuda_launcher'}, {'test_file': 'export/test_dynamic_shapes'}, {'test_file': 'dynamo/test_sdpa'}, {'test_file': 'inductor/test_aot_inductor_package'}, {'test_file': 'dynamo/test_utils'}, {'test_file': 'inductor/test_provenance_tracing'}, {'test_file': 'inductor/test_combo_kernels'}, {'test_file': 'inductor/test_codegen_triton'}, {'test_file': 'dynamo/test_frame_init'}, {'test_file': 'inductor/test_device_assert'}, {'test_file': 'dynamo/test_skip_non_tensor'}, {'test_file': 'dynamo/test_skip_guard_eval_unsafe'}, {'test_file': 'inductor/test_decompose_mem_bound_mm'}, {'test_file': 'dynamo/test_interop'}, {'test_file': 'functorch/test_eager_transforms'}, {'test_file': 'inductor/test_control_deps'}, {'test_file': 'inductor/test_benchmarking'}, {'test_file': 'inductor/test_helion_kernels'}, {'test_file': 'inductor/test_quantization'}, {'test_file': 'inductor/test_op_dtype_prop'}, {'test_file': 'inductor/test_best_config'}, {'test_file': 'export/test_tools'}, {'test_file': 'inductor/test_compiled_optimizers'}, {'test_file': 'dynamo/test_buffers_override'}, {'test_file': 'inductor/test_inplacing_pass'}, {'test_file': 'inductor/test_aot_inductor_custom_ops'}, {'test_file': 'inductor/test_split_cat_fx_passes'}, {'test_file': 'inductor/test_profiler'}, {'test_file': 'inductor/test_memory_planning'}, {'test_file': 'inductor/test_mem_estimation'}, {'test_file': 'dynamo/test_view'}, {'test_file': 'inductor/test_cutlass_evt'}, {'test_file': 'dynamo/test_reconstruct'}, {'test_file': 'dynamo/test_aot_autograd'}, {'test_file': 'export/test_cpp_serdes'}, {'test_file': 'inductor/test_cache'}, {'test_file': 'inductor/test_block_analysis'}, {'test_file': 'inductor/test_online_softmax'}, {'test_file': 'dynamo/test_subgraphs'}, {'test_file': 'dynamo/test_pre_dispatch'}, {'test_file': 'inductor/test_custom_post_grad_passes'}, {'test_file': 'dynamo/test_fx_annotate'}, {'test_file': 'dynamo/test_pgo'}, {'test_file': 'dynamo/test_config'}, {'test_file': 'dynamo/test_metrics_context'}, {'test_file': 'export/test_package'}, {'test_file': 'export/test_export_opinfo'}, {'test_file': 'dynamo/test_nops'}, {'test_file': 'inductor/test_graph_transform_observer'}, {'test_file': 'inductor/test_memory'}, {'test_file': 'inductor/test_aot_inductor_utils'}, {'test_file': 'export/test_db'}, {'test_file': 'dynamo/test_export_mutations'}, {'test_file': 'inductor/test_config'}, {'test_file': 'inductor/test_dependencies'}, {'test_file': 'inductor/test_fuzzer'}, {'test_file': 'dynamo/test_global'}, {'test_file': 'inductor/test_control_flow'}, {'test_file': 'dynamo/test_graph_region_tracker'}, {'test_file': 'dynamo/test_unittest'}, {'test_file': 'dynamo/test_streams'}, {'test_file': 'inductor/test_compile'}, {'test_file': 'inductor/test_unbacked_symints'}, {'test_file': 'inductor/test_scatter_optimization'}, {'test_file': 'dynamo/test_functions'}, {'test_file': 'inductor/test_mix_order_reduction'}, {'test_file': 'inductor/test_ordered_set'}, {'test_file': 'inductor/test_pallas'}, {'test_file': 'dynamo/test_install_free_tensors'}, {'test_file': 'inductor/test_torchinductor_codegen_config_overrides'}, {'test_file': 'export/test_passes'}, {'test_file': 'dynamo/test_autograd_function'}, {'test_file': 'inductor/test_codecache'}, {'test_file': 'dynamo/test_cudagraphs'}, {'test_file': 'inductor/test_alignment'}, {'test_file': 'inductor/test_padding'}, {'test_file': 'dynamo/test_profiler'}, {'test_file': 'dynamo/test_guard_serialization'}, {'test_file': 'dynamo/test_compile'}, {'test_file': 'dynamo/test_nested_graph_breaks'}, {'test_file': 'dynamo/test_dicts'}, {'test_file': 'inductor/test_needs_exact_strides'}, {'test_file': 'inductor/test_auto_functionalize'}, {'test_file': 'inductor/test_split_cat_fx_aten_passes'}, {'test_file': 'inductor/test_minifier_isolate'}, {'test_file': 'dynamo/test_aot_compile'}, {'test_file': 'dynamo/test_list'}, {'test_file': 'dynamo/test_resume'}, {'test_file': 'inductor/test_augmented_graph_helper'}, {'test_file': 'dynamo/test_deviceguard'}, {'test_file': 'dynamo/test_sources'}, {'test_file': 'dynamo/test_backward_higher_order_ops'}, {'test_file': 'dynamo/test_modes'}, {'test_file': 'dynamo/test_optimizers'}, {'test_file': 'export/test_torchbind'}, {'test_file': 'inductor/test_custom_partitioner_fn'}, {'test_file': 'dynamo/test_debug_utils'}, {'test_file': 'dynamo/test_base_hop'}, {'test_file': 'dynamo/test_export'}, {'test_file': 'dynamo/test_sets'}, {'test_file': 'dynamo/test_package'}, {'test_file': 'inductor/test_efficient_conv_bn_eval'}, {'test_file': 'inductor/test_torchbind'}, {'test_file': 'dynamo/test_python_dispatcher'}, {'test_file': 'export/test_swap'}, {'test_file': 'export/test_unflatten'}, {'test_file': 'dynamo/test_verify_correctness'}, {'test_file': 'dynamo/test_wrap_inductor_compiled_regions'}, {'test_file': 'inductor/test_fxir_backend'}, {'test_file': 'dynamo/test_callback'}, {'test_file': 'dynamo/test_cudagraphs_expandable_segments'}, {'test_file': 'inductor/test_caching'}, {'test_file': 'dynamo/test_aot_autograd_cache'}, {'test_file': 'dynamo/test_flat_apply'}, {'test_file': 'dynamo/test_input_attr_tracking'}, {'test_file': 'dynamo/test_compiler_bisector'}, {'test_file': 'dynamo/test_graph_deduplication'}, {'test_file': 'inductor/test_distributed_patterns'}, {'test_file': 'dynamo/test_structured_trace'}, {'test_file': 'dynamo/test_error_messages'}, {'test_file': 'dynamo/test_bytecode_utils'}, {'test_file': 'dynamo/test_fake_distributed'}, {'test_file': 'inductor/test_mps_basic'}, {'test_file': 'export/test_nativert'}, {'test_file': 'inductor/test_custom_op_autotune'}, {'test_file': 'export/test_hop'}, {'test_file': 'dynamo/test_tree_map'}, {'test_file': 'dynamo/test_minifier'}, {'test_file': 'dynamo/test_guard_manager'}, {'test_file': 'export/test_schema'}, {'test_file': 'dynamo/test_torchrec'}, {'test_file': 'export/test_pass_infra'}, {'test_file': 'dynamo/test_recompile_ux'}, {'test_file': 'inductor/test_cudagraph_trees_expandable_segments'}, {'test_file': 'export/test_experimental'}, {'test_file': 'export/test_converter'}, {'test_file': 'export/test_export'}, {'test_file': 'test_model_exports_to_core_aten'}, {'test_file': 'dynamo/test_precompile_context'}, {'test_file': 'dynamo/test_trace_rules'}, {'test_file': 'export/test_upgrader'}, {'test_file': 'dynamo/test_hooks'}, {'test_file': 'dynamo/test_reorder_logs'}, {'test_file': 'dynamo/test_subclasses'}, {'test_file': 'dynamo/test_exceptions'}, {'test_file': 'dynamo/test_generator'}, {'test_file': 'export/test_lift_unlift'}, {'test_file': 'dynamo/test_decorators'}, {'test_file': 'export/test_verifier'}, {'test_file': 'export/test_sparse'}, {'test_file': 'dynamo/test_python_autograd'}, {'test_file': 'export/test_draft_export'}, {'test_file': 'dynamo/test_comptime'}, {'test_file': 'test_sort_and_select'}, {'test_file': 'test_privateuseone_python_backend'}, {'test_file': 'functorch/test_rearrange'}, {'test_file': 'functorch/test_parsing'}, {'test_file': 'test_varlen_attention'}, {'test_file': 'test_package'}, {'test_file': 'test_cuda'}, {'test_file': 'profiler/test_profiler'}, {'test_file': 'test_mkl_verbose'}, {'test_file': 'test_comparison_utils'}, {'test_file': 'functorch/test_ac_logging'}, {'test_file': 'test_mkldnn_verbose'}, {'test_file': 'test_cpp_api_parity'}, {'test_file': 'test_utils_config_module'}, {'test_file': 'test_hop_infra'}, {'test_file': 'test_transformers'}, {'test_file': 'test_appending_byte_serializer'}, {'test_file': 'test_license'}, {'test_file': 'test_ao_sparsity'}, {'test_file': 'test_autoload'}, {'test_file': 'nn/attention/test_open_registry'}, {'test_file': 'xpu/test_fusion'}, {'test_file': 'test_as_strided'}, {'test_file': 'test_foreach'}, {'test_file': 'test_proxy_tensor'}, {'test_file': 'torch_np/test_binary_ufuncs'}, {'test_file': 'torch_np/test_unary_ufuncs'}, {'test_file': 'test_utils_filelock'}, {'test_file': 'test_extension_utils'}, {'test_file': 'test_rename_privateuse1_to_existing_device'}, {'test_file': 'nn/attention/test_fa4'}, {'test_file': 'typing/test_python_operators'}, {'test_file': 'test_functionalization'}, {'test_file': 'test_matmul_cuda'}, {'test_file': 'torch_np/test_dtype'}, {'test_file': 'test_file_check'}, {'test_file': 'profiler/test_kineto'}, {'test_file': 'test_flop_counter'}, {'test_file': 'backends/xeon/test_launch'}, {'test_file': 'test_show_pickle'}, {'test_file': 'test_openmp'}, {'test_file': 'test_expanded_weights'}, {'test_file': 'test_module_tracker'}, {'test_file': 'test_autograd'}, {'test_file': 'torch_np/numpy_tests/core/test_scalarinherit'}, {'test_file': 'test_tensorexpr_pybind'}, {'test_file': 'test_fx_experimental'}, {'test_file': 'functorch/test_ac_knapsack'}, {'test_file': 'torch_np/test_nep50_examples'}, {'test_file': 'test_torch'}, {'test_file': 'xpu/test_gemm'}, {'test_file': 'test_sparse'}, {'test_file': 'test_fx_passes'}, {'test_file': 'functorch/test_logging'}, {'test_file': 'higher_order_ops/test_local_map'}, {'test_file': 'test_namedtensor'}, {'test_file': 'test_tensorexpr'}, {'test_file': 'test_dataloader'}, {'test_file': 'functorch/test_minifier'}, {'test_file': 'higher_order_ops/test_invoke_quant'}, {'test_file': 'torch_np/test_basic'}, {'test_file': 'test_jiterator'}, {'test_file': 'test_native_functions'}, {'test_file': 'test_typing'}, {'test_file': 'higher_order_ops/test_with_effects'}, {'test_file': 'higher_order_ops/test_invoke_subgraph'}, {'test_file': 'test_weak'}, {'test_file': 'test_decomp'}, {'test_file': 'test_complex'}, {'test_file': 'test_optim'}, {'test_file': 'lazy/test_functionalization'}, {'test_file': 'torch_np/test_random'}, {'test_file': 'nn/test_multihead_attention'}, {'test_file': 'test_legacy_vmap'}, {'test_file': 'lazy/test_bindings'}, {'test_file': 'xpu/test_conv'}, {'test_file': 'test_utils'}, {'test_file': 'test_pytree'}, {'test_file': 'test_namedtuple_return_api'}, {'test_file': 'test_ci_sanity_check_fail'}, {'test_file': 'profiler/test_record_function'}, {'test_file': 'test_compile_benchmark_util'}, {'test_file': 'test_set_default_mobile_cpu_allocator'}, {'test_file': 'test_fake_tensor'}, {'test_file': 'test_stateless'}, {'test_file': 'functorch/test_ac'}, {'test_file': 'test_binary_ufuncs'}, {'test_file': 'higher_order_ops/test_print'}, {'test_file': 'test_per_overload_api'}, {'test_file': 'torch_np/numpy_tests/core/test_einsum'}, {'test_file': 'test_multiprocessing'}, {'test_file': 'test_out_dtype_op'}, {'test_file': 'torch_np/test_ufuncs_basic'}, {'test_file': 'lazy/test_step_closures'}, {'test_file': 'functorch/dim/test_getsetitem'}, {'test_file': 'test_ops_fwd_gradients'}, {'test_file': 'test_meta'}, {'test_file': 'test_ops_jit'}, {'test_file': 'test_fx'}, {'test_file': 'test_numpy_interop'}, {'test_file': 'profiler/test_cpp_thread'}, {'test_file': 'test_ops_gradients'}, {'test_file': 'test_hub'}, {'test_file': 'test_segment_reductions'}, {'test_file': 'test_opaque_obj_v2'}, {'test_file': 'test_autograd_fallback'}, {'test_file': 'test_type_hints'}, {'test_file': 'functorch/test_aot_joint_with_descriptors'}, {'test_file': 'test_nestedtensor'}, {'test_file': 'distributions/test_constraints'}, {'test_file': 'test_linalg'}, {'test_file': 'test_functionalization_of_rng_ops'}, {'test_file': 'test_fx_reinplace_pass'}, {'test_file': 'functorch/test_control_flow'}, {'test_file': 'test_cuda_expandable_segments'}, {'test_file': 'test_modules'}, {'test_file': 'nn/test_packed_sequence'}, {'test_file': 'test_numa_binding'}, {'test_file': 'test_pruning_op'}, {'test_file': 'test_jit_fuser_te'}, {'test_file': 'test_autocast'}, {'test_file': 'test_logging'}, {'test_file': 'test_python_dispatch'}, {'test_file': 'nn/test_lazy_modules'}, {'test_file': 'nn/test_pruning'}, {'test_file': 'test_monitor'}, {'test_file': 'test_cuda_sanitizer'}, {'test_file': 'test_bundled_inputs'}, {'test_file': 'torch_np/numpy_tests/core/test_numeric'}, {'test_file': 'torch_np/numpy_tests/core/test_multiarray'}, {'test_file': 'test_itt'}, {'test_file': 'torch_np/numpy_tests/lib/test_function_base'}, {'test_file': 'test_masked'}, {'test_file': 'test_sympy_utils'}, {'test_file': 'test_jit_disabled'}, {'test_file': 'test_subclass'}, {'test_file': 'test_import_stats'}, {'test_file': 'functorch/test_vmap_registrations'}, {'test_file': 'nn/test_parametrization'}, {'test_file': 'test_mkldnn_fusion'}, {'test_file': 'test_cpp_extensions_mtia_backend'}, {'test_file': 'lazy/test_ts_opinfo'}, {'test_file': 'test_dynamic_shapes'}, {'test_file': 'complex_tensor/test_complex_tensor'}, {'test_file': 'optim/test_lrscheduler'}, {'test_file': 'optim/test_swa_utils'}, {'test_file': 'cpp_extensions/python_agnostic_extension/test/test_python_agnostic'}, {'test_file': 'functorch/test_memory_efficient_fusion'}, {'test_file': 'torch_np/numpy_tests/lib/test_histograms'}, {'test_file': 'torch_np/test_indexing'}, {'test_file': 'test_schema_check'}, {'test_file': 'test_tensorboard'}, {'test_file': 'torch_np/numpy_tests/core/test_indexing'}, {'test_file': 'test_futures'}, {'test_file': 'test_tensor_creation_ops'}, {'test_file': 'nn/test_dropout'}, {'test_file': 'functorch/dim/test_split'}, {'test_file': 'torch_np/numpy_tests/lib/test_type_check'}, {'test_file': 'cpp_extensions/test_libtorch_agnostic'}, {'test_file': 'test_cpp_extensions_stream_and_event'}, {'test_file': 'profiler/test_execution_trace'}, {'test_file': 'test_jit'}, {'test_file': 'test_dispatch'}, {'test_file': 'test_datapipe'}, {'test_file': 'test_numba_integration'}, {'test_file': 'test_functional_optim'}, {'test_file': 'test_maskedtensor'}, {'test_file': 'test_ops'}, {'test_file': 'benchmark_utils/test_benchmark_utils'}, {'test_file': 'functorch/test_dims'}, {'test_file': 'torch_np/numpy_tests/core/test_scalarmath'}, {'test_file': 'test_scaled_matmul_cuda'}, {'test_file': 'torch_np/numpy_tests/core/test_shape_base'}, {'test_file': 'test_vulkan'}, {'test_file': 'lazy/test_generator'}, {'test_file': 'nn/test_convolution'}, {'test_file': 'torch_np/numpy_tests/linalg/test_linalg'}, {'test_file': 'torch_np/numpy_tests/core/test_dtype'}, {'test_file': 'lazy/test_debug_util'}, {'test_file': 'nn/test_load_state_dict'}, {'test_file': 'test_shape_ops'}, {'test_file': 'nn/test_module_hooks'}, {'test_file': 'torch_np/numpy_tests/lib/test_twodim_base'}, {'test_file': 'profiler/test_memory_profiler'}, {'test_file': 'test_jit_llga_fuser'}, {'test_file': 'test_serialization'}, {'test_file': 'test_sparse_csr'}, {'test_file': 'optim/test_optim'}, {'test_file': 'test_indexing'}, {'test_file': 'torch_np/numpy_tests/fft/test_pocketfft'}, {'test_file': 'functorch/test_ops'}, {'test_file': 'torch_np/numpy_tests/lib/test_shape_base_'}, {'test_file': 'test_cpp_extensions_jit'}, {'test_file': 'torch_np/numpy_tests/core/test_getlimits'}, {'test_file': 'torch_np/test_ndarray_methods'}, {'test_file': 'test_view_ops'}, {'test_file': 'test_type_info'}, {'test_file': 'functorch/test_aotdispatch'}, {'test_file': 'test_nn'}, {'test_file': 'torch_np/numpy_tests/core/test_dlpack'}, {'test_file': 'test_multiprocessing_spawn'}, {'test_file': 'test_scatter_gather_ops'}, {'test_file': 'test_cuda_multigpu'}, {'test_file': 'test_mkldnn'}, {'test_file': 'functorch/test_vmap'}, {'test_file': 'torch_np/numpy_tests/lib/test_index_tricks'}, {'test_file': 'test_jit_autocast'}, {'test_file': 'nn/test_pooling'}, {'test_file': 'nn/test_embedding'}, {'test_file': 'test_unary_ufuncs'}, {'test_file': 'test_xnnpack_integration'}, {'test_file': 'test_cuda_trace'}, {'test_file': 'test_native_mha'}, {'test_file': 'torch_np/numpy_tests/core/test_numerictypes'}, {'test_file': 'test_function_schema'}, {'test_file': 'test_accelerator'}, {'test_file': 'nn/test_init'}, {'test_file': 'torch_np/numpy_tests/core/test_scalar_methods'}, {'test_file': 'torch_np/numpy_tests/fft/test_helper'}, {'test_file': 'test_mobile_optimizer'}, {'test_file': 'test_overrides'}, {'test_file': 'torch_np/test_function_base'}, {'test_file': 'test_type_promotion'}, {'test_file': 'torch_np/test_scalars_0D_arrays'}, {'test_file': 'test_cuda_primary_ctx'}, {'test_file': 'profiler/test_profiler_tree'}, {'test_file': 'torch_np/numpy_tests/lib/test_arraysetops'}, {'test_file': 'test_dlpack'}, {'test_file': 'profiler/test_torch_tidy'}, {'test_file': 'lazy/test_reuse_ir'}, {'test_file': 'test_functional_autograd_benchmark'}, {'test_file': 'test_reductions'}, {'test_file': 'torch_np/test_reductions'}, {'test_file': 'torch_np/numpy_tests/core/test_scalar_ctors'}, {'test_file': 'torch_np/numpy_tests/lib/test_arraypad'}, {'test_file': 'test_prims'}, {'test_file': 'test_spectral_ops'}, {'test_file': 'profiler/test_python_tracer'}, {'test_file': 'cpp_extensions/libtorch_agnostic_2_10_extension/test_version_compatibility'}, {'test_file': 'test_torchfuzz_repros'}, {'test_file': 'distributions/test_distributions'}, {'test_file': 'doctests'}, {'test_file': 'test_autoload_disable'}, {'test_file': 'test_autoload_enable'}, {'test_file': 'test_cpp_extensions_aot_ninja'}, {'test_file': 'test_cpp_extensions_aot_no_ninja'}], 'excluded': []} from test/test-reports/td_exclusions-fb2b987622427d634578.json is not a benchmark record, skipping 2025-12-04T10:17:36.3129200Z warn(f"{result} from {filepath} is not a benchmark record, skipping") 2025-12-04T10:17:36.3189128Z Prepare all required actions 2025-12-04T10:17:36.3189424Z Getting action download info 2025-12-04T10:17:36.3202712Z ##[group]Run ./.github/actions/teardown-rocm 2025-12-04T10:17:36.3202865Z env: 2025-12-04T10:17:36.3202988Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:36.3203154Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:36.3203380Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:36.3203579Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:36.3203982Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:36.3204371Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:36.3204513Z AWS_REGION: us-east-1 2025-12-04T10:17:36.3204715Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:36.3204917Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:36.3207221Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:36.3207713Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:36.3207913Z DEVICE_NAME: rocm 2025-12-04T10:17:36.3208043Z DEVICE_TYPE: AMD Instinct MI325X 2025-12-04T10:17:36.3208187Z ##[endgroup] 2025-12-04T10:17:36.3215077Z ##[group]Run # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T10:17:36.3215325Z # ignore expansion of "docker ps -q" since it could be empty 2025-12-04T10:17:36.3215513Z # shellcheck disable=SC2046 2025-12-04T10:17:36.3215679Z docker stop $(docker ps -q) || true 2025-12-04T10:17:36.3215835Z # Prune all stopped containers. 2025-12-04T10:17:36.3215991Z docker container prune -f 2025-12-04T10:17:36.3219857Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:17:36.3220028Z env: 2025-12-04T10:17:36.3220150Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:36.3220323Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:36.3220532Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:36.3220716Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:36.3221119Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:36.3221510Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:36.3221646Z AWS_REGION: us-east-1 2025-12-04T10:17:36.3221793Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:36.3221968Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:36.3224234Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:36.3224426Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:36.3224624Z DEVICE_NAME: rocm 2025-12-04T10:17:36.3224749Z DEVICE_TYPE: AMD Instinct MI325X 2025-12-04T10:17:36.3224889Z ##[endgroup] 2025-12-04T10:17:47.0879875Z eda51ac9bc2e 2025-12-04T10:17:49.5887699Z Deleted Containers: 2025-12-04T10:17:49.5888153Z eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:49.5888406Z 2025-12-04T10:17:49.5888511Z Total reclaimed space: 10.32GB 2025-12-04T10:17:49.5925545Z Prepare all required actions 2025-12-04T10:17:49.5941185Z ##[group]Run ./.github/actions/diskspace-cleanup 2025-12-04T10:17:49.5941341Z with: 2025-12-04T10:17:49.5941560Z diskspace-cutoff: 70 2025-12-04T10:17:49.5941686Z env: 2025-12-04T10:17:49.5941796Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:49.5941955Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:49.5942155Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:49.5942346Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:49.5942921Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:49.5943320Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:49.5943457Z AWS_REGION: us-east-1 2025-12-04T10:17:49.5943654Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:49.5943845Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:49.5946134Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:49.5946326Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:49.5946527Z DEVICE_NAME: rocm 2025-12-04T10:17:49.5946654Z DEVICE_TYPE: AMD Instinct MI325X 2025-12-04T10:17:49.5946843Z ##[endgroup] 2025-12-04T10:17:49.5967142Z ##[group]Run set -ex 2025-12-04T10:17:49.5967300Z set -ex 2025-12-04T10:17:49.5967423Z diskspace_cutoff=70 2025-12-04T10:17:49.5967605Z docker_root_dir=$(docker info -f '{{.DockerRootDir}}') 2025-12-04T10:17:49.5967791Z if [ ! -d "$docker_root_dir" ]; then 2025-12-04T10:17:49.5968017Z  echo "Docker root directory ($docker_root_dir) does not exist. Skipping disk space check." 2025-12-04T10:17:49.5968237Z  exit 0 2025-12-04T10:17:49.5968355Z fi 2025-12-04T10:17:49.5968545Z diskspace=$(df -H --output=pcent ${docker_root_dir} | sed -n 2p | sed 's/%//' | sed 's/ //') 2025-12-04T10:17:49.5968901Z msg="Please file an issue on pytorch/pytorch reporting the faulty runner. Include a link to the runner logs so the runner can be identified" 2025-12-04T10:17:49.5969215Z if [[ "$diskspace" -ge "$diskspace_cutoff" ]] ; then 2025-12-04T10:17:49.5969388Z  docker system prune -af 2025-12-04T10:17:49.5969605Z  diskspace_new=$(df -H --output=pcent ${docker_root_dir} | sed -n 2p | sed 's/%//' | sed 's/ //') 2025-12-04T10:17:49.5969850Z  if [[ "$diskspace_new" -gt "$diskspace_cutoff" ]] ; then 2025-12-04T10:17:49.5970047Z  diskspace_cutoff_int=$((diskspace_cutoff + 0)) 2025-12-04T10:17:49.5970234Z  difference=$((100 - diskspace_cutoff_int)) 2025-12-04T10:17:49.5970474Z  echo "Error: Available diskspace is less than $difference percent. Not enough diskspace." 2025-12-04T10:17:49.5970689Z  echo "$msg" 2025-12-04T10:17:49.5970825Z  exit 1 2025-12-04T10:17:49.5970949Z  else 2025-12-04T10:17:49.5971099Z  difference=$((diskspace - diskspace_new)) 2025-12-04T10:17:49.5971290Z  echo "Diskspace saved: $difference percent" 2025-12-04T10:17:49.5971444Z  fi 2025-12-04T10:17:49.5971562Z fi 2025-12-04T10:17:49.5975628Z shell: /usr/bin/bash --noprofile --norc -e -o pipefail {0} 2025-12-04T10:17:49.5975797Z env: 2025-12-04T10:17:49.5975917Z GIT_DEFAULT_BRANCH: main 2025-12-04T10:17:49.5976081Z RUNNER_ARTIFACT_DIR: /home/runner/_work/_temp/artifacts 2025-12-04T10:17:49.5976281Z RUNNER_TEST_RESULTS_DIR: /home/runner/_work/_temp/test-results 2025-12-04T10:17:49.5976477Z RUNNER_DOCS_DIR: /home/runner/_work/_temp/docs 2025-12-04T10:17:49.5976950Z GPU_FLAG: --device=/dev/mem --device=/dev/kfd --group-add 110 --device /dev/dri/renderD152 --group-add video --group-add 109 --group-add daemon --group-add bin --cap-add=SYS_PTRACE --security-opt seccomp=unconfined --network=host 2025-12-04T10:17:49.5977425Z AWS_DEFAULT_REGION: us-east-1 2025-12-04T10:17:49.5977569Z AWS_REGION: us-east-1 2025-12-04T10:17:49.5977731Z AWS_ACCESS_KEY_ID: *** 2025-12-04T10:17:49.5977942Z AWS_SECRET_ACCESS_KEY: *** 2025-12-04T10:17:49.5980230Z AWS_SESSION_TOKEN: *** 2025-12-04T10:17:49.5980427Z CONTAINER_NAME: eda51ac9bc2e38470ba980220f8bdfa9240ea7ab11182239105521c84ef615fa 2025-12-04T10:17:49.5980633Z DEVICE_NAME: rocm 2025-12-04T10:17:49.5980766Z DEVICE_TYPE: AMD Instinct MI325X 2025-12-04T10:17:49.5980907Z ##[endgroup] 2025-12-04T10:17:49.5999581Z + diskspace_cutoff=70 2025-12-04T10:17:49.6002223Z ++ docker info -f '{{.DockerRootDir}}' 2025-12-04T10:17:49.6325947Z + docker_root_dir=/home/runner/docker-data 2025-12-04T10:17:49.6326204Z + '[' '!' -d /home/runner/docker-data ']' 2025-12-04T10:17:49.6332906Z ++ df -H --output=pcent /home/runner/docker-data 2025-12-04T10:17:49.6333394Z ++ sed -n 2p 2025-12-04T10:17:49.6334642Z ++ sed s/%// 2025-12-04T10:17:49.6335158Z ++ sed 's/ //' 2025-12-04T10:17:49.6349159Z + diskspace=' 4' 2025-12-04T10:17:49.6349682Z + msg='Please file an issue on pytorch/pytorch reporting the faulty runner. Include a link to the runner logs so the runner can be identified' 2025-12-04T10:17:49.6350108Z + [[ 4 -ge 70 ]] 2025-12-04T10:17:49.6379216Z Post job cleanup. 2025-12-04T10:17:49.6401075Z Post job cleanup. 2025-12-04T10:17:49.7002619Z Post job cleanup. 2025-12-04T10:17:49.7014459Z Post job cleanup. 2025-12-04T10:17:49.7309663Z Logging out of registry 308535385114.dkr.ecr.us-east-1.amazonaws.com 2025-12-04T10:17:49.7477281Z Post job cleanup. 2025-12-04T10:17:49.8080176Z Post job cleanup. 2025-12-04T10:17:49.8114995Z Post job cleanup. 2025-12-04T10:17:49.8565655Z [command]/usr/bin/git version 2025-12-04T10:17:49.8593881Z git version 2.52.0 2025-12-04T10:17:49.8616058Z Copying '/home/runner/.gitconfig' to '/home/runner/_work/_temp/fc32abb9-5fc9-4b7b-b568-ea9d6f0c966e/.gitconfig' 2025-12-04T10:17:49.8622211Z Temporarily overriding HOME='/home/runner/_work/_temp/fc32abb9-5fc9-4b7b-b568-ea9d6f0c966e' before making global git config changes 2025-12-04T10:17:49.8622801Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T10:17:49.8624813Z [command]/usr/bin/git config --global --add safe.directory /home/runner/_work/pytorch/pytorch 2025-12-04T10:17:49.8652800Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T10:17:49.8669415Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T10:17:49.8846266Z Entering 'android/libs/fbjni' 2025-12-04T10:17:49.8881107Z Entering 'third_party/FP16' 2025-12-04T10:17:49.8905479Z Entering 'third_party/FXdiv' 2025-12-04T10:17:49.8926524Z Entering 'third_party/NNPACK' 2025-12-04T10:17:49.8952558Z Entering 'third_party/NVTX' 2025-12-04T10:17:49.8978152Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:17:49.8998476Z Entering 'third_party/XNNPACK' 2025-12-04T10:17:49.9024191Z Entering 'third_party/aiter' 2025-12-04T10:17:49.9046212Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:17:49.9074911Z Entering 'third_party/benchmark' 2025-12-04T10:17:49.9102663Z Entering 'third_party/composable_kernel' 2025-12-04T10:17:49.9129398Z Entering 'third_party/cpp-httplib' 2025-12-04T10:17:49.9151388Z Entering 'third_party/cpuinfo' 2025-12-04T10:17:49.9172809Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:17:49.9197088Z Entering 'third_party/cutlass' 2025-12-04T10:17:49.9221608Z Entering 'third_party/fbgemm' 2025-12-04T10:17:49.9245415Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:17:49.9265987Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:17:49.9292056Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:17:49.9311524Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:17:49.9341298Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:17:49.9362745Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:17:49.9382866Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:17:49.9404186Z Entering 'third_party/flash-attention' 2025-12-04T10:17:49.9431733Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:17:49.9458236Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:17:49.9485335Z Entering 'third_party/flatbuffers' 2025-12-04T10:17:49.9516373Z Entering 'third_party/fmt' 2025-12-04T10:17:49.9538259Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:17:49.9562512Z Entering 'third_party/gloo' 2025-12-04T10:17:49.9584413Z Entering 'third_party/googletest' 2025-12-04T10:17:49.9606665Z Entering 'third_party/ideep' 2025-12-04T10:17:49.9630410Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:17:49.9655904Z Entering 'third_party/ittapi' 2025-12-04T10:17:49.9684420Z Entering 'third_party/kineto' 2025-12-04T10:17:49.9707989Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:17:49.9732235Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:17:49.9754099Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:17:49.9775292Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:17:49.9796401Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:17:49.9830096Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:17:49.9851982Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:17:49.9872744Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:17:49.9894228Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:17:49.9921447Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:17:49.9942698Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:17:49.9966367Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:49.9987736Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:50.0011181Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:17:50.0030387Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:17:50.0052234Z Entering 'third_party/kleidiai' 2025-12-04T10:17:50.0074175Z Entering 'third_party/mimalloc' 2025-12-04T10:17:50.0095138Z Entering 'third_party/nlohmann' 2025-12-04T10:17:50.0118087Z Entering 'third_party/onnx' 2025-12-04T10:17:50.0144242Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:17:50.0174327Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:17:50.0197066Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:17:50.0217188Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:17:50.0242384Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:17:50.0263295Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:17:50.0282082Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:17:50.0300157Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:17:50.0320516Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:17:50.0345921Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:50.0379424Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:50.0405880Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:17:50.0445578Z Entering 'third_party/pocketfft' 2025-12-04T10:17:50.0468847Z Entering 'third_party/protobuf' 2025-12-04T10:17:50.0494993Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:17:50.0526072Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:17:50.0554271Z Entering 'third_party/psimd' 2025-12-04T10:17:50.0578244Z Entering 'third_party/pthreadpool' 2025-12-04T10:17:50.0607308Z Entering 'third_party/pybind11' 2025-12-04T10:17:50.0633499Z Entering 'third_party/python-peachpy' 2025-12-04T10:17:50.0655733Z Entering 'third_party/sleef' 2025-12-04T10:17:50.0677821Z Entering 'third_party/tensorpipe' 2025-12-04T10:17:50.0708694Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:17:50.0735345Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:17:50.0761154Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:17:50.0784192Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:17:50.0808963Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:17:50.0858525Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T10:17:50.0874624Z http.https://github.com/.extraheader 2025-12-04T10:17:50.0882245Z [command]/usr/bin/git config --local --unset-all http.https://github.com/.extraheader 2025-12-04T10:17:50.0902803Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T10:17:50.1058871Z Entering 'android/libs/fbjni' 2025-12-04T10:17:50.1072446Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1090913Z Entering 'third_party/FP16' 2025-12-04T10:17:50.1104507Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1126594Z Entering 'third_party/FXdiv' 2025-12-04T10:17:50.1140701Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1162188Z Entering 'third_party/NNPACK' 2025-12-04T10:17:50.1178836Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1196725Z Entering 'third_party/NVTX' 2025-12-04T10:17:50.1218931Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1243309Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:17:50.1257095Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1273997Z Entering 'third_party/XNNPACK' 2025-12-04T10:17:50.1288134Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1313730Z Entering 'third_party/aiter' 2025-12-04T10:17:50.1325604Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1351274Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:17:50.1365513Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1394651Z Entering 'third_party/benchmark' 2025-12-04T10:17:50.1409506Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1429013Z Entering 'third_party/composable_kernel' 2025-12-04T10:17:50.1442873Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1464223Z Entering 'third_party/cpp-httplib' 2025-12-04T10:17:50.1478243Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1496861Z Entering 'third_party/cpuinfo' 2025-12-04T10:17:50.1510045Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1530167Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:17:50.1543613Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1561693Z Entering 'third_party/cutlass' 2025-12-04T10:17:50.1574438Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1594998Z Entering 'third_party/fbgemm' 2025-12-04T10:17:50.1607919Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1627394Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:17:50.1647195Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1663095Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:17:50.1678655Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1700619Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:17:50.1716618Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1736151Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:17:50.1750714Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1772550Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:17:50.1789723Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1806700Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:17:50.1821154Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1839517Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:17:50.1856261Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1874657Z Entering 'third_party/flash-attention' 2025-12-04T10:17:50.1888268Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1904909Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:17:50.1916996Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1935898Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:17:50.1951656Z http.https://github.com/.extraheader 2025-12-04T10:17:50.1977343Z Entering 'third_party/flatbuffers' 2025-12-04T10:17:50.1991152Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2011985Z Entering 'third_party/fmt' 2025-12-04T10:17:50.2024568Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2043929Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:17:50.2057041Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2074268Z Entering 'third_party/gloo' 2025-12-04T10:17:50.2086626Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2104602Z Entering 'third_party/googletest' 2025-12-04T10:17:50.2116340Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2135608Z Entering 'third_party/ideep' 2025-12-04T10:17:50.2148260Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2164952Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:17:50.2179621Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2202150Z Entering 'third_party/ittapi' 2025-12-04T10:17:50.2218644Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2236343Z Entering 'third_party/kineto' 2025-12-04T10:17:50.2254036Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2272536Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:17:50.2288564Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2306572Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:17:50.2322801Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2341594Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:17:50.2357356Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2378142Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:17:50.2401265Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2418778Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:17:50.2444608Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2469929Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:17:50.2485836Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2506974Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:17:50.2524258Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2550123Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:17:50.2567404Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2586015Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:17:50.2606846Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2625739Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:17:50.2640802Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2667125Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:17:50.2685896Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2710951Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:50.2727221Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2749948Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:50.2765500Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2790649Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:17:50.2804602Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2827957Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:17:50.2841809Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2866354Z Entering 'third_party/kleidiai' 2025-12-04T10:17:50.2883450Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2903091Z Entering 'third_party/mimalloc' 2025-12-04T10:17:50.2918336Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2936557Z Entering 'third_party/nlohmann' 2025-12-04T10:17:50.2955787Z http.https://github.com/.extraheader 2025-12-04T10:17:50.2972393Z Entering 'third_party/onnx' 2025-12-04T10:17:50.2989689Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3014280Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:17:50.3029834Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3053679Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:17:50.3068375Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3093103Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:17:50.3110028Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3128926Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:17:50.3144535Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3164159Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:17:50.3177806Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3202493Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:17:50.3215419Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3235879Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:17:50.3252469Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3271906Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:17:50.3287819Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3309335Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:17:50.3323931Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3343489Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:50.3358629Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3376615Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:50.3391478Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3411109Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:17:50.3425278Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3453611Z Entering 'third_party/pocketfft' 2025-12-04T10:17:50.3470360Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3490089Z Entering 'third_party/protobuf' 2025-12-04T10:17:50.3508240Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3529641Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:17:50.3551020Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3567816Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:17:50.3590456Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3614424Z Entering 'third_party/psimd' 2025-12-04T10:17:50.3635722Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3654203Z Entering 'third_party/pthreadpool' 2025-12-04T10:17:50.3669321Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3687977Z Entering 'third_party/pybind11' 2025-12-04T10:17:50.3702931Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3722617Z Entering 'third_party/python-peachpy' 2025-12-04T10:17:50.3737573Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3757864Z Entering 'third_party/sleef' 2025-12-04T10:17:50.3772002Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3788869Z Entering 'third_party/tensorpipe' 2025-12-04T10:17:50.3802251Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3821623Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:17:50.3835952Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3853390Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:17:50.3866404Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3884086Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:17:50.3897190Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3915286Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:17:50.3929570Z http.https://github.com/.extraheader 2025-12-04T10:17:50.3945817Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:17:50.3962664Z http.https://github.com/.extraheader 2025-12-04T10:17:50.4000783Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.4025230Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T10:17:50.4190526Z Entering 'android/libs/fbjni' 2025-12-04T10:17:50.4202549Z file:/home/runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T10:17:50.4215987Z Entering 'third_party/FP16' 2025-12-04T10:17:50.4230870Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T10:17:50.4238634Z Entering 'third_party/FXdiv' 2025-12-04T10:17:50.4257344Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T10:17:50.4268203Z Entering 'third_party/NNPACK' 2025-12-04T10:17:50.4281732Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T10:17:50.4291707Z Entering 'third_party/NVTX' 2025-12-04T10:17:50.4303371Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T10:17:50.4318568Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:17:50.4331917Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T10:17:50.4344217Z Entering 'third_party/XNNPACK' 2025-12-04T10:17:50.4360103Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T10:17:50.4393981Z Entering 'third_party/aiter' 2025-12-04T10:17:50.4406941Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T10:17:50.4418091Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:17:50.4428959Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T10:17:50.4442979Z Entering 'third_party/benchmark' 2025-12-04T10:17:50.4454222Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:17:50.4464496Z Entering 'third_party/composable_kernel' 2025-12-04T10:17:50.4475961Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T10:17:50.4488890Z Entering 'third_party/cpp-httplib' 2025-12-04T10:17:50.4501312Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T10:17:50.4510967Z Entering 'third_party/cpuinfo' 2025-12-04T10:17:50.4526270Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T10:17:50.4535771Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:17:50.4547085Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T10:17:50.4561205Z Entering 'third_party/cutlass' 2025-12-04T10:17:50.4572219Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T10:17:50.4584428Z Entering 'third_party/fbgemm' 2025-12-04T10:17:50.4596432Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T10:17:50.4607584Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:17:50.4619781Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T10:17:50.4629231Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:17:50.4638705Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T10:17:50.4650432Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:17:50.4659951Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T10:17:50.4667903Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:17:50.4677111Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T10:17:50.4689398Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:17:50.4700310Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T10:17:50.4708551Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:17:50.4717778Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T10:17:50.4725399Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:17:50.4735596Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T10:17:50.4745172Z Entering 'third_party/flash-attention' 2025-12-04T10:17:50.4756726Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T10:17:50.4766072Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:17:50.4774721Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T10:17:50.4786422Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:17:50.4796059Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T10:17:50.4808166Z Entering 'third_party/flatbuffers' 2025-12-04T10:17:50.4818289Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T10:17:50.4828128Z Entering 'third_party/fmt' 2025-12-04T10:17:50.4838222Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T10:17:50.4846835Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:17:50.4858006Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T10:17:50.4867117Z Entering 'third_party/gloo' 2025-12-04T10:17:50.4878273Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T10:17:50.4887495Z Entering 'third_party/googletest' 2025-12-04T10:17:50.4897738Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:17:50.4906813Z Entering 'third_party/ideep' 2025-12-04T10:17:50.4917179Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T10:17:50.4927494Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:17:50.4936893Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T10:17:50.4948712Z Entering 'third_party/ittapi' 2025-12-04T10:17:50.4958929Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T10:17:50.4968093Z Entering 'third_party/kineto' 2025-12-04T10:17:50.4978968Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T10:17:50.4988307Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:17:50.4997617Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T10:17:50.5006016Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:17:50.5014860Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T10:17:50.5025253Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:17:50.5034896Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T10:17:50.5043752Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:17:50.5054194Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T10:17:50.5063213Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:17:50.5072627Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T10:17:50.5081360Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:17:50.5091934Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T10:17:50.5101657Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:17:50.5110853Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T10:17:50.5119118Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:17:50.5128647Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:17:50.5137563Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:17:50.5146945Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T10:17:50.5155785Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:17:50.5164716Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T10:17:50.5173320Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:17:50.5182905Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T10:17:50.5191825Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:50.5201540Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T10:17:50.5210254Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:50.5219702Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T10:17:50.5230371Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:17:50.5243462Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T10:17:50.5252693Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:17:50.5265129Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T10:17:50.5276014Z Entering 'third_party/kleidiai' 2025-12-04T10:17:50.5286151Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T10:17:50.5295445Z Entering 'third_party/mimalloc' 2025-12-04T10:17:50.5305445Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T10:17:50.5314885Z Entering 'third_party/nlohmann' 2025-12-04T10:17:50.5325277Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T10:17:50.5334686Z Entering 'third_party/onnx' 2025-12-04T10:17:50.5345762Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T10:17:50.5363298Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:17:50.5374307Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:17:50.5386025Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:17:50.5397104Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T10:17:50.5407135Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:17:50.5419384Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:17:50.5428257Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:17:50.5438349Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:17:50.5447210Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:17:50.5457981Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T10:17:50.5466725Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:17:50.5476186Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T10:17:50.5484487Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:17:50.5494700Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T10:17:50.5502231Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:17:50.5512062Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T10:17:50.5520330Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:17:50.5529933Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T10:17:50.5538308Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:50.5548107Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T10:17:50.5557438Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:50.5567651Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T10:17:50.5577685Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:17:50.5587033Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T10:17:50.5605515Z Entering 'third_party/pocketfft' 2025-12-04T10:17:50.5615901Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T10:17:50.5625286Z Entering 'third_party/protobuf' 2025-12-04T10:17:50.5635881Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T10:17:50.5646445Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:17:50.5655076Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:17:50.5664199Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:17:50.5673323Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:17:50.5684288Z Entering 'third_party/psimd' 2025-12-04T10:17:50.5694730Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T10:17:50.5703809Z Entering 'third_party/pthreadpool' 2025-12-04T10:17:50.5714745Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T10:17:50.5723705Z Entering 'third_party/pybind11' 2025-12-04T10:17:50.5735496Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:17:50.5745314Z Entering 'third_party/python-peachpy' 2025-12-04T10:17:50.5760424Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T10:17:50.5771730Z Entering 'third_party/sleef' 2025-12-04T10:17:50.5791635Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T10:17:50.5801732Z Entering 'third_party/tensorpipe' 2025-12-04T10:17:50.5812486Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T10:17:50.5822751Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:17:50.5836423Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:17:50.5845646Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:17:50.5856804Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T10:17:50.5866657Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:17:50.5879349Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T10:17:50.5889519Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:17:50.5911114Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:17:50.5919319Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:17:50.5933203Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T10:17:50.5961966Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.5980421Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.5999550Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6016061Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6032787Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6049670Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6064217Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6078257Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6092239Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6106251Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6121615Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6135758Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6152044Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6165870Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6188271Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6203231Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6217606Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6231469Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6247333Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6261560Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6275389Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6295818Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6310883Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6324831Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6337394Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6352237Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6366680Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6380483Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6395461Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6409174Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6422987Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6436868Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6450386Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6464270Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6478970Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6494137Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6509399Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6523959Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6539203Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6557582Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6572715Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6591467Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6607006Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6622283Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6636609Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6652244Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6666395Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6679679Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6693107Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6707085Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6720819Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6734686Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6749733Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6764114Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6779257Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6793544Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6808369Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6823046Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6838467Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6853133Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6867643Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6883894Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6898575Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6913129Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6928596Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6942906Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6957386Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6971988Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.6987084Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7001324Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7015494Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7029078Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7042235Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7058903Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7073682Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7089219Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7104448Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7119455Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7134344Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7147850Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7162591Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:50.7263347Z Post job cleanup. 2025-12-04T10:17:50.7709951Z [command]/usr/bin/git version 2025-12-04T10:17:50.7737512Z git version 2.52.0 2025-12-04T10:17:50.7758010Z Copying '/home/runner/.gitconfig' to '/home/runner/_work/_temp/53a77264-5803-49fb-a0f3-b0241573a131/.gitconfig' 2025-12-04T10:17:50.7763582Z Temporarily overriding HOME='/home/runner/_work/_temp/53a77264-5803-49fb-a0f3-b0241573a131' before making global git config changes 2025-12-04T10:17:50.7763928Z Adding repository directory to the temporary git global config as a safe directory 2025-12-04T10:17:50.7766186Z [command]/usr/bin/git config --global --add safe.directory /home/runner/_work/pytorch/pytorch 2025-12-04T10:17:50.7795993Z [command]/usr/bin/git config --local --name-only --get-regexp core\.sshCommand 2025-12-04T10:17:50.7814168Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2025-12-04T10:17:50.7991455Z Entering 'android/libs/fbjni' 2025-12-04T10:17:50.8015317Z Entering 'third_party/FP16' 2025-12-04T10:17:50.8038822Z Entering 'third_party/FXdiv' 2025-12-04T10:17:50.8063306Z Entering 'third_party/NNPACK' 2025-12-04T10:17:50.8086841Z Entering 'third_party/NVTX' 2025-12-04T10:17:50.8107279Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:17:50.8129841Z Entering 'third_party/XNNPACK' 2025-12-04T10:17:50.8156819Z Entering 'third_party/aiter' 2025-12-04T10:17:50.8181366Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:17:50.8213255Z Entering 'third_party/benchmark' 2025-12-04T10:17:50.8234437Z Entering 'third_party/composable_kernel' 2025-12-04T10:17:50.8260911Z Entering 'third_party/cpp-httplib' 2025-12-04T10:17:50.8286975Z Entering 'third_party/cpuinfo' 2025-12-04T10:17:50.8310637Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:17:50.8334751Z Entering 'third_party/cutlass' 2025-12-04T10:17:50.8360335Z Entering 'third_party/fbgemm' 2025-12-04T10:17:50.8389999Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:17:50.8412876Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:17:50.8438212Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:17:50.8465901Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:17:50.8490266Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:17:50.8512792Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:17:50.8533216Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:17:50.8557329Z Entering 'third_party/flash-attention' 2025-12-04T10:17:50.8580376Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:17:50.8602155Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:17:50.8626316Z Entering 'third_party/flatbuffers' 2025-12-04T10:17:50.8648180Z Entering 'third_party/fmt' 2025-12-04T10:17:50.8674372Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:17:50.8696521Z Entering 'third_party/gloo' 2025-12-04T10:17:50.8739927Z Entering 'third_party/googletest' 2025-12-04T10:17:50.8762927Z Entering 'third_party/ideep' 2025-12-04T10:17:50.8784962Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:17:50.8808998Z Entering 'third_party/ittapi' 2025-12-04T10:17:50.8831610Z Entering 'third_party/kineto' 2025-12-04T10:17:50.8859518Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:17:50.8880999Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:17:50.8903070Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:17:50.8927039Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:17:50.8947838Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:17:50.8967749Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:17:50.8991706Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:17:50.9017416Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:17:50.9038971Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:17:50.9064255Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:17:50.9094397Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:17:50.9122421Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:50.9145680Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:50.9173151Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:17:50.9193881Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:17:50.9221697Z Entering 'third_party/kleidiai' 2025-12-04T10:17:50.9246610Z Entering 'third_party/mimalloc' 2025-12-04T10:17:50.9267269Z Entering 'third_party/nlohmann' 2025-12-04T10:17:50.9288359Z Entering 'third_party/onnx' 2025-12-04T10:17:50.9317570Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:17:50.9340439Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:17:50.9362231Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:17:50.9383262Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:17:50.9406130Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:17:50.9429563Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:17:50.9449876Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:17:50.9473935Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:17:50.9495302Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:17:50.9516159Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:50.9543817Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:50.9569172Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:17:50.9602597Z Entering 'third_party/pocketfft' 2025-12-04T10:17:50.9629997Z Entering 'third_party/protobuf' 2025-12-04T10:17:50.9660091Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:17:50.9682389Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:17:50.9712542Z Entering 'third_party/psimd' 2025-12-04T10:17:50.9736322Z Entering 'third_party/pthreadpool' 2025-12-04T10:17:50.9759319Z Entering 'third_party/pybind11' 2025-12-04T10:17:50.9780945Z Entering 'third_party/python-peachpy' 2025-12-04T10:17:50.9805646Z Entering 'third_party/sleef' 2025-12-04T10:17:50.9828750Z Entering 'third_party/tensorpipe' 2025-12-04T10:17:50.9852551Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:17:50.9872507Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:17:50.9896918Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:17:50.9920418Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:17:50.9939671Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:17:50.9987982Z [command]/usr/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2025-12-04T10:17:51.0007430Z [command]/usr/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2025-12-04T10:17:51.0162484Z Entering 'android/libs/fbjni' 2025-12-04T10:17:51.0187596Z Entering 'third_party/FP16' 2025-12-04T10:17:51.0210592Z Entering 'third_party/FXdiv' 2025-12-04T10:17:51.0233321Z Entering 'third_party/NNPACK' 2025-12-04T10:17:51.0259324Z Entering 'third_party/NVTX' 2025-12-04T10:17:51.0280278Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:17:51.0302059Z Entering 'third_party/XNNPACK' 2025-12-04T10:17:51.0328530Z Entering 'third_party/aiter' 2025-12-04T10:17:51.0352826Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:17:51.0379003Z Entering 'third_party/benchmark' 2025-12-04T10:17:51.0400780Z Entering 'third_party/composable_kernel' 2025-12-04T10:17:51.0427103Z Entering 'third_party/cpp-httplib' 2025-12-04T10:17:51.0449786Z Entering 'third_party/cpuinfo' 2025-12-04T10:17:51.0470978Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:17:51.0491893Z Entering 'third_party/cutlass' 2025-12-04T10:17:51.0519765Z Entering 'third_party/fbgemm' 2025-12-04T10:17:51.0543215Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:17:51.0563896Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:17:51.0586475Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:17:51.0607083Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:17:51.0645828Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:17:51.0667267Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:17:51.0692098Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:17:51.0718570Z Entering 'third_party/flash-attention' 2025-12-04T10:17:51.0743898Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:17:51.0768136Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:17:51.0808037Z Entering 'third_party/flatbuffers' 2025-12-04T10:17:51.0831671Z Entering 'third_party/fmt' 2025-12-04T10:17:51.0855036Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:17:51.0875443Z Entering 'third_party/gloo' 2025-12-04T10:17:51.0896578Z Entering 'third_party/googletest' 2025-12-04T10:17:51.0917637Z Entering 'third_party/ideep' 2025-12-04T10:17:51.0943548Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:17:51.0970347Z Entering 'third_party/ittapi' 2025-12-04T10:17:51.0994789Z Entering 'third_party/kineto' 2025-12-04T10:17:51.1017493Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:17:51.1038017Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:17:51.1060726Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:17:51.1081253Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:17:51.1100562Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:17:51.1125384Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:17:51.1151720Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:17:51.1173292Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:17:51.1195680Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:17:51.1217360Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:17:51.1242569Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:17:51.1264133Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:51.1285717Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:51.1310925Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:17:51.1330235Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:17:51.1351587Z Entering 'third_party/kleidiai' 2025-12-04T10:17:51.1377380Z Entering 'third_party/mimalloc' 2025-12-04T10:17:51.1399505Z Entering 'third_party/nlohmann' 2025-12-04T10:17:51.1421199Z Entering 'third_party/onnx' 2025-12-04T10:17:51.1455615Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:17:51.1480668Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:17:51.1502901Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:17:51.1523920Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:17:51.1545821Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:17:51.1568051Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:17:51.1590365Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:17:51.1618323Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:17:51.1640208Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:17:51.1661837Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:51.1687578Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:51.1709702Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:17:51.1738937Z Entering 'third_party/pocketfft' 2025-12-04T10:17:51.1760270Z Entering 'third_party/protobuf' 2025-12-04T10:17:51.1789725Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:17:51.1813695Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:17:51.1840084Z Entering 'third_party/psimd' 2025-12-04T10:17:51.1862445Z Entering 'third_party/pthreadpool' 2025-12-04T10:17:51.1884064Z Entering 'third_party/pybind11' 2025-12-04T10:17:51.1905833Z Entering 'third_party/python-peachpy' 2025-12-04T10:17:51.1931459Z Entering 'third_party/sleef' 2025-12-04T10:17:51.1956285Z Entering 'third_party/tensorpipe' 2025-12-04T10:17:51.1981008Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:17:51.2002825Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:17:51.2024745Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:17:51.2044968Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:17:51.2081515Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:17:51.2135746Z [command]/usr/bin/git config --local --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.2157450Z [command]/usr/bin/git submodule foreach --recursive git config --local --show-origin --name-only --get-regexp remote.origin.url 2025-12-04T10:17:51.2327121Z Entering 'android/libs/fbjni' 2025-12-04T10:17:51.2338080Z file:/home/runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config remote.origin.url 2025-12-04T10:17:51.2349849Z Entering 'third_party/FP16' 2025-12-04T10:17:51.2370380Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config remote.origin.url 2025-12-04T10:17:51.2384775Z Entering 'third_party/FXdiv' 2025-12-04T10:17:51.2398537Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config remote.origin.url 2025-12-04T10:17:51.2410541Z Entering 'third_party/NNPACK' 2025-12-04T10:17:51.2421129Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config remote.origin.url 2025-12-04T10:17:51.2432348Z Entering 'third_party/NVTX' 2025-12-04T10:17:51.2442349Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config remote.origin.url 2025-12-04T10:17:51.2453527Z Entering 'third_party/VulkanMemoryAllocator' 2025-12-04T10:17:51.2471372Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config remote.origin.url 2025-12-04T10:17:51.2473235Z Entering 'third_party/XNNPACK' 2025-12-04T10:17:51.2495941Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config remote.origin.url 2025-12-04T10:17:51.2497925Z Entering 'third_party/aiter' 2025-12-04T10:17:51.2508586Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config remote.origin.url 2025-12-04T10:17:51.2516525Z Entering 'third_party/aiter/3rdparty/composable_kernel' 2025-12-04T10:17:51.2529211Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config remote.origin.url 2025-12-04T10:17:51.2538106Z Entering 'third_party/benchmark' 2025-12-04T10:17:51.2555861Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:17:51.2557780Z Entering 'third_party/composable_kernel' 2025-12-04T10:17:51.2568723Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config remote.origin.url 2025-12-04T10:17:51.2579975Z Entering 'third_party/cpp-httplib' 2025-12-04T10:17:51.2593108Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config remote.origin.url 2025-12-04T10:17:51.2599957Z Entering 'third_party/cpuinfo' 2025-12-04T10:17:51.2616689Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config remote.origin.url 2025-12-04T10:17:51.2625626Z Entering 'third_party/cudnn_frontend' 2025-12-04T10:17:51.2638714Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config remote.origin.url 2025-12-04T10:17:51.2649076Z Entering 'third_party/cutlass' 2025-12-04T10:17:51.2660816Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config remote.origin.url 2025-12-04T10:17:51.2675667Z Entering 'third_party/fbgemm' 2025-12-04T10:17:51.2688340Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config remote.origin.url 2025-12-04T10:17:51.2701872Z Entering 'third_party/fbgemm/external/asmjit' 2025-12-04T10:17:51.2715793Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config remote.origin.url 2025-12-04T10:17:51.2728305Z Entering 'third_party/fbgemm/external/composable_kernel' 2025-12-04T10:17:51.2740286Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config remote.origin.url 2025-12-04T10:17:51.2753406Z Entering 'third_party/fbgemm/external/cpuinfo' 2025-12-04T10:17:51.2765373Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config remote.origin.url 2025-12-04T10:17:51.2778389Z Entering 'third_party/fbgemm/external/cutlass' 2025-12-04T10:17:51.2791076Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config remote.origin.url 2025-12-04T10:17:51.2804239Z Entering 'third_party/fbgemm/external/googletest' 2025-12-04T10:17:51.2818466Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config remote.origin.url 2025-12-04T10:17:51.2828795Z Entering 'third_party/fbgemm/external/hipify_torch' 2025-12-04T10:17:51.2841601Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config remote.origin.url 2025-12-04T10:17:51.2852778Z Entering 'third_party/fbgemm/external/json' 2025-12-04T10:17:51.2867048Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config remote.origin.url 2025-12-04T10:17:51.2876963Z Entering 'third_party/flash-attention' 2025-12-04T10:17:51.2891010Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config remote.origin.url 2025-12-04T10:17:51.2905317Z Entering 'third_party/flash-attention/csrc/composable_kernel' 2025-12-04T10:17:51.2919697Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config remote.origin.url 2025-12-04T10:17:51.2938693Z Entering 'third_party/flash-attention/csrc/cutlass' 2025-12-04T10:17:51.2951252Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config remote.origin.url 2025-12-04T10:17:51.2967625Z Entering 'third_party/flatbuffers' 2025-12-04T10:17:51.2980621Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config remote.origin.url 2025-12-04T10:17:51.2992345Z Entering 'third_party/fmt' 2025-12-04T10:17:51.3003411Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config remote.origin.url 2025-12-04T10:17:51.3012627Z Entering 'third_party/gemmlowp/gemmlowp' 2025-12-04T10:17:51.3024505Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config remote.origin.url 2025-12-04T10:17:51.3035416Z Entering 'third_party/gloo' 2025-12-04T10:17:51.3048023Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config remote.origin.url 2025-12-04T10:17:51.3056790Z Entering 'third_party/googletest' 2025-12-04T10:17:51.3066994Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:17:51.3074732Z Entering 'third_party/ideep' 2025-12-04T10:17:51.3083666Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config remote.origin.url 2025-12-04T10:17:51.3092235Z Entering 'third_party/ideep/mkl-dnn' 2025-12-04T10:17:51.3099246Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config remote.origin.url 2025-12-04T10:17:51.3111133Z Entering 'third_party/ittapi' 2025-12-04T10:17:51.3118794Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config remote.origin.url 2025-12-04T10:17:51.3127823Z Entering 'third_party/kineto' 2025-12-04T10:17:51.3136333Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config remote.origin.url 2025-12-04T10:17:51.3144508Z Entering 'third_party/kineto/libkineto/third_party/dynolog' 2025-12-04T10:17:51.3151571Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config remote.origin.url 2025-12-04T10:17:51.3158282Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/DCGM' 2025-12-04T10:17:51.3167298Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config remote.origin.url 2025-12-04T10:17:51.3177343Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/cpr' 2025-12-04T10:17:51.3188125Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config remote.origin.url 2025-12-04T10:17:51.3196091Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/fmt' 2025-12-04T10:17:51.3205316Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config remote.origin.url 2025-12-04T10:17:51.3213365Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags' 2025-12-04T10:17:51.3231357Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config remote.origin.url 2025-12-04T10:17:51.3241078Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/gflags/doc' 2025-12-04T10:17:51.3249801Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config remote.origin.url 2025-12-04T10:17:51.3259537Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/glog' 2025-12-04T10:17:51.3269768Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config remote.origin.url 2025-12-04T10:17:51.3277649Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/googletest' 2025-12-04T10:17:51.3289511Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:17:51.3297421Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/json' 2025-12-04T10:17:51.3307201Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config remote.origin.url 2025-12-04T10:17:51.3315103Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/pfs' 2025-12-04T10:17:51.3323820Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config remote.origin.url 2025-12-04T10:17:51.3331971Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp' 2025-12-04T10:17:51.3341135Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T10:17:51.3350318Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:51.3361091Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T10:17:51.3371353Z Entering 'third_party/kineto/libkineto/third_party/dynolog/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:51.3381166Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T10:17:51.3393135Z Entering 'third_party/kineto/libkineto/third_party/fmt' 2025-12-04T10:17:51.3402696Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config remote.origin.url 2025-12-04T10:17:51.3410928Z Entering 'third_party/kineto/libkineto/third_party/googletest' 2025-12-04T10:17:51.3420865Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config remote.origin.url 2025-12-04T10:17:51.3430703Z Entering 'third_party/kleidiai' 2025-12-04T10:17:51.3442658Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config remote.origin.url 2025-12-04T10:17:51.3451655Z Entering 'third_party/mimalloc' 2025-12-04T10:17:51.3461126Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config remote.origin.url 2025-12-04T10:17:51.3469957Z Entering 'third_party/nlohmann' 2025-12-04T10:17:51.3479285Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config remote.origin.url 2025-12-04T10:17:51.3488741Z Entering 'third_party/onnx' 2025-12-04T10:17:51.3498066Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config remote.origin.url 2025-12-04T10:17:51.3513593Z Entering 'third_party/onnx/third_party/pybind11' 2025-12-04T10:17:51.3524299Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:17:51.3536702Z Entering 'third_party/opentelemetry-cpp' 2025-12-04T10:17:51.3548809Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config remote.origin.url 2025-12-04T10:17:51.3558526Z Entering 'third_party/opentelemetry-cpp/third_party/benchmark' 2025-12-04T10:17:51.3575086Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:17:51.3584672Z Entering 'third_party/opentelemetry-cpp/third_party/googletest' 2025-12-04T10:17:51.3594665Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:17:51.3603291Z Entering 'third_party/opentelemetry-cpp/third_party/ms-gsl' 2025-12-04T10:17:51.3612288Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config remote.origin.url 2025-12-04T10:17:51.3620585Z Entering 'third_party/opentelemetry-cpp/third_party/nlohmann-json' 2025-12-04T10:17:51.3629982Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config remote.origin.url 2025-12-04T10:17:51.3639284Z Entering 'third_party/opentelemetry-cpp/third_party/opentelemetry-proto' 2025-12-04T10:17:51.3648463Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config remote.origin.url 2025-12-04T10:17:51.3657516Z Entering 'third_party/opentelemetry-cpp/third_party/opentracing-cpp' 2025-12-04T10:17:51.3667869Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config remote.origin.url 2025-12-04T10:17:51.3676178Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp' 2025-12-04T10:17:51.3685832Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config remote.origin.url 2025-12-04T10:17:51.3694059Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/civetweb' 2025-12-04T10:17:51.3704131Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config remote.origin.url 2025-12-04T10:17:51.3712534Z Entering 'third_party/opentelemetry-cpp/third_party/prometheus-cpp/3rdparty/googletest' 2025-12-04T10:17:51.3721757Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config remote.origin.url 2025-12-04T10:17:51.3733411Z Entering 'third_party/opentelemetry-cpp/tools/vcpkg' 2025-12-04T10:17:51.3742487Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config remote.origin.url 2025-12-04T10:17:51.3759564Z Entering 'third_party/pocketfft' 2025-12-04T10:17:51.3774413Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config remote.origin.url 2025-12-04T10:17:51.3785493Z Entering 'third_party/protobuf' 2025-12-04T10:17:51.3803107Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config remote.origin.url 2025-12-04T10:17:51.3820976Z Entering 'third_party/protobuf/third_party/benchmark' 2025-12-04T10:17:51.3830556Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config remote.origin.url 2025-12-04T10:17:51.3840579Z Entering 'third_party/protobuf/third_party/googletest' 2025-12-04T10:17:51.3852890Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:17:51.3866474Z Entering 'third_party/psimd' 2025-12-04T10:17:51.3877387Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config remote.origin.url 2025-12-04T10:17:51.3886635Z Entering 'third_party/pthreadpool' 2025-12-04T10:17:51.3896090Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config remote.origin.url 2025-12-04T10:17:51.3905569Z Entering 'third_party/pybind11' 2025-12-04T10:17:51.3915129Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:17:51.3923735Z Entering 'third_party/python-peachpy' 2025-12-04T10:17:51.3933860Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config remote.origin.url 2025-12-04T10:17:51.3942422Z Entering 'third_party/sleef' 2025-12-04T10:17:51.3953300Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config remote.origin.url 2025-12-04T10:17:51.3961676Z Entering 'third_party/tensorpipe' 2025-12-04T10:17:51.3971080Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config remote.origin.url 2025-12-04T10:17:51.3979753Z Entering 'third_party/tensorpipe/third_party/googletest' 2025-12-04T10:17:51.3988958Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config remote.origin.url 2025-12-04T10:17:51.3997739Z Entering 'third_party/tensorpipe/third_party/libnop' 2025-12-04T10:17:51.4008274Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config remote.origin.url 2025-12-04T10:17:51.4017503Z Entering 'third_party/tensorpipe/third_party/libuv' 2025-12-04T10:17:51.4026789Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config remote.origin.url 2025-12-04T10:17:51.4036087Z Entering 'third_party/tensorpipe/third_party/pybind11' 2025-12-04T10:17:51.4046071Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config remote.origin.url 2025-12-04T10:17:51.4054587Z Entering 'third_party/tensorpipe/third_party/pybind11/tools/clang' 2025-12-04T10:17:51.4064475Z file:/home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config remote.origin.url 2025-12-04T10:17:51.4090505Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/android/libs/fbjni/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4109287Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FP16/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4125723Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/FXdiv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4139163Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4158015Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NVTX/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4174481Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/VulkanMemoryAllocator/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4190802Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/XNNPACK/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4205346Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4220106Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/aiter/modules/3rdparty/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4234133Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4250525Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4265538Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpp-httplib/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4281442Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4295375Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cudnn_frontend/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4310645Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4326188Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4341180Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/asmjit/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4354870Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4369761Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cpuinfo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4383190Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4397252Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4412596Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/hipify_torch/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4427001Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fbgemm/modules/external/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4440996Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4454738Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/composable_kernel/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4469039Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flash-attention/modules/csrc/cutlass/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4483982Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/flatbuffers/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4498646Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4512092Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/gemmlowp/gemmlowp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4525968Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/gloo/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4540409Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4554809Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4570626Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ideep/modules/mkl-dnn/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4584272Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/ittapi/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4597527Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4612505Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4627832Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/DCGM/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4641648Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/cpr/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4656414Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4671654Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4686635Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/gflags/modules/doc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4701938Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/glog/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4718635Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4733988Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4749246Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/pfs/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4764894Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4779849Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4793900Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/dynolog/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4808032Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/fmt/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4822157Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kineto/modules/libkineto/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4836189Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/kleidiai/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4849889Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/mimalloc/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4862398Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/nlohmann/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4876873Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4891595Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/onnx/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4906471Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4921297Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4934742Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4948468Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/ms-gsl/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4964267Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/nlohmann-json/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4979518Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentelemetry-proto/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.4993712Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/opentracing-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5008785Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5027116Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/civetweb/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5042254Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/third_party/prometheus-cpp/modules/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5061548Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/opentelemetry-cpp/modules/tools/vcpkg/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5076267Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/pocketfft/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5090449Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5105818Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/benchmark/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5120760Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/protobuf/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5135059Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/psimd/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5149484Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/NNPACK_deps/pthreadpool/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5164228Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5179219Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/python-peachpy/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5193501Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/sleef/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5207288Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5220541Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/googletest/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5233678Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libnop/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5249019Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/libuv/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5263287Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5276374Z [command]/usr/bin/git config --file /home/runner/_work/pytorch/pytorch/.git/modules/third_party/tensorpipe/modules/third_party/pybind11/modules/tools/clang/config --name-only --get-regexp ^includeIf\.gitdir: 2025-12-04T10:17:51.5372100Z Cleaning up orphan processes